+ echo Logging output to /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//log_node10.txt Logging output to /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//log_node10.txt + export ASCEND_PROCESS_LOG_PATH=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/10 + ASCEND_PROCESS_LOG_PATH=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/10 + mkdir -p /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/10 + DATA_PATH=/local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml + TOKENIZER_PATH=/data_4/models/Qwen/Qwen2.5-14B-Instruct/ + CKPT_LOAD_DIR=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/ + VIT_CKPT_LOAD_DIR=/ + CKPT_SAVE_DIR=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743// + rsync -avh /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743/ sending incremental file list sent 71 bytes received 12 bytes 166.00 bytes/sec total size is 23.84K speedup is 287.17 + cd /local_disk/cognitron_vl/ + rm -fr datasets + mkdir -p datasets + ln -s /data/data/ datasets/CV + ln -s /data/data/LLM datasets/LLM + ln -s /data/data/LMM datasets/LMM + source /local_disk/cognitron_vl//scripts/set_env_mg_npu.sh ++ source /usr/local/Ascend/driver/bin/setenv.bash +++ DEP_INFO_FILE=/etc/ascend_install.info +++ [[ -f /etc/ascend_install.info ]] +++ . /etc/ascend_install.info +++ DRV_LIB64_COMMON_LDPATH=/driver/lib64/common +++ DRV_LIB64_DRV_LDPATH=/driver/lib64/driver +++ DRV_LIB64_LDPATH=/driver/lib64 +++ export LD_LIBRARY_PATH=/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ LD_LIBRARY_PATH=/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ export PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin +++ PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin ++ source /usr/local/Ascend/ascend-toolkit/set_env.sh +++ export LD_LIBRARY_PATH=/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ LD_LIBRARY_PATH=/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ export ASCEND_TOOLKIT_HOME=/usr/local/Ascend/ascend-toolkit/latest +++ ASCEND_TOOLKIT_HOME=/usr/local/Ascend/ascend-toolkit/latest ++++ arch +++ export LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ export LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial: +++ export PYTHONPATH=/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe: +++ PYTHONPATH=/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe: +++ export PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin +++ PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin +++ export ASCEND_AICPU_PATH=/usr/local/Ascend/ascend-toolkit/latest +++ ASCEND_AICPU_PATH=/usr/local/Ascend/ascend-toolkit/latest +++ export ASCEND_OPP_PATH=/usr/local/Ascend/ascend-toolkit/latest/opp +++ ASCEND_OPP_PATH=/usr/local/Ascend/ascend-toolkit/latest/opp +++ export TOOLCHAIN_HOME=/usr/local/Ascend/ascend-toolkit/latest/toolkit +++ TOOLCHAIN_HOME=/usr/local/Ascend/ascend-toolkit/latest/toolkit +++ export ASCEND_HOME_PATH=/usr/local/Ascend/ascend-toolkit/latest +++ ASCEND_HOME_PATH=/usr/local/Ascend/ascend-toolkit/latest ++ export HCCL_CONNECT_TIMEOUT=7200 ++ HCCL_CONNECT_TIMEOUT=7200 ++ export HCCL_EXEC_TIMEOUT=7200 ++ HCCL_EXEC_TIMEOUT=7200 ++ export COMBINED_ENABLE=1 ++ COMBINED_ENABLE=1 ++ export MULTI_STREAM_MEMORY_REUSE=1 ++ MULTI_STREAM_MEMORY_REUSE=1 ++ export HCCL_RDMA_TC=160 ++ HCCL_RDMA_TC=160 ++ export HCCL_RDMA_SL=5 ++ HCCL_RDMA_SL=5 ++ export HCCL_INTRA_PCIE_ENABLE=0 ++ HCCL_INTRA_PCIE_ENABLE=0 ++ export HCCL_INTRA_ROCE_ENABLE=1 ++ HCCL_INTRA_ROCE_ENABLE=1 ++ export HCCL_RDMA_TIMEOUT=20 ++ HCCL_RDMA_TIMEOUT=20 ++ export INF_NAN_MODE_ENABLE=1 ++ INF_NAN_MODE_ENABLE=1 ++ export DISTRIBUTED_BACKEND=hccl ++ DISTRIBUTED_BACKEND=hccl ++ export ASCEND_LAUNCH_BLOCKING=0 ++ ASCEND_LAUNCH_BLOCKING=0 ++ export ASCEND_SLOG_PRINT_TO_STDOUT=0 ++ ASCEND_SLOG_PRINT_TO_STDOUT=0 ++ export ASCEND_GLOBAL_LOG_LEVEL=3 ++ ASCEND_GLOBAL_LOG_LEVEL=3 ++ export ASCEND_GLOBAL_EVENT_ENABLE=0 ++ ASCEND_GLOBAL_EVENT_ENABLE=0 ++ export TASK_QUEUE_ENABLE=1 ++ TASK_QUEUE_ENABLE=1 ++ export PTCOPY_ENABLE=1 ++ PTCOPY_ENABLE=1 ++ export COMBINED_ENABLE=1 ++ COMBINED_ENABLE=1 ++ export DYNAMIC_OP=ADD#MUL ++ DYNAMIC_OP=ADD#MUL ++ export HCCL_WHITELIST_DISABLE=1 ++ HCCL_WHITELIST_DISABLE=1 ++ export HCCL_CONNECT_TIMEOUT=7200 ++ HCCL_CONNECT_TIMEOUT=7200 ++ export HCCL_WHITELIST_DISABLE=1 ++ HCCL_WHITELIST_DISABLE=1 ++ export CUDA_DEVICE_MAX_CONNECTIONS=1 ++ CUDA_DEVICE_MAX_CONNECTIONS=1 ++ pip3 install --no-index --find-links=/data/software/ -r requirements_npu.txt Looking in links: /data/software/ Processing data/software/expecttest-0.2.1-py3-none-any.whl (from -r requirements_npu.txt (line 1)) Requirement already satisfied: peft in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 2)) (0.7.0) Processing data/software/XlsxWriter-3.2.0-py3-none-any.whl (from -r requirements_npu.txt (line 3)) Requirement already satisfied: termcolor in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 4)) (2.4.0) Requirement already satisfied: tabulate in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 5)) (0.9.0) Processing data/software/tiktoken-0.7.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from -r requirements_npu.txt (line 6)) Requirement already satisfied: matplotlib in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 7)) (3.7.5) Processing data/software/datasets-3.0.0-py3-none-any.whl (from -r requirements_npu.txt (line 8)) Requirement already satisfied: einops in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 9)) (0.7.0) Processing data/software/pybind11-2.13.6-py3-none-any.whl (from -r requirements_npu.txt (line 10)) Requirement already satisfied: tensorboardX in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 11)) (2.6.2.2) Processing data/software/pyarrow-17.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from -r requirements_npu.txt (line 12)) Requirement already satisfied: transformers>=4.40.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 13)) (4.40.1) Requirement already satisfied: deepspeed>=0.14.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 14)) (0.14.5) Processing data/software/accelerate-0.34.2-py3-none-any.whl (from -r requirements_npu.txt (line 15)) Requirement already satisfied: timm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 16)) (0.9.16) Processing data/software/flask-3.0.3-py3-none-any.whl (from -r requirements_npu.txt (line 17)) Processing data/software/Flask_RESTful-0.3.10-py2.py3-none-any.whl (from -r requirements_npu.txt (line 18)) Processing data/software/decord-0.6.0-py3-none-manylinux2010_x86_64.whl (from -r requirements_npu.txt (line 19)) Processing data/software/natsort-8.4.0-py3-none-any.whl (from -r requirements_npu.txt (line 20)) Requirement already satisfied: numpy>=1.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (1.24.4) Requirement already satisfied: packaging>=20.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (23.2) Requirement already satisfied: psutil in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (5.9.8) Requirement already satisfied: pyyaml in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (5.4.1) Requirement already satisfied: torch>=1.13.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (2.1.0+cpu) Requirement already satisfied: tqdm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (4.66.2) Requirement already satisfied: safetensors in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (0.4.2) Requirement already satisfied: huggingface-hub>=0.17.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (0.20.3) Requirement already satisfied: regex>=2022.1.18 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tiktoken->-r requirements_npu.txt (line 6)) (2023.12.25) Requirement already satisfied: requests>=2.26.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tiktoken->-r requirements_npu.txt (line 6)) (2.31.0) Requirement already satisfied: contourpy>=1.0.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (1.1.1) Requirement already satisfied: cycler>=0.10 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (0.12.1) Requirement already satisfied: fonttools>=4.22.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (4.49.0) Requirement already satisfied: kiwisolver>=1.0.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (1.4.5) Requirement already satisfied: pillow>=6.2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (10.2.0) Requirement already satisfied: pyparsing>=2.3.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (3.1.1) Requirement already satisfied: python-dateutil>=2.7 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (2.8.2) Requirement already satisfied: importlib-resources>=3.2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (6.1.2) Requirement already satisfied: filelock in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.13.1) Requirement already satisfied: dill<0.3.9,>=0.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (0.3.7) Requirement already satisfied: pandas in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2.0.3) Processing data/software/requests-2.32.3-py3-none-any.whl (from tiktoken->-r requirements_npu.txt (line 6)) Processing data/software/tqdm-4.67.1-py3-none-any.whl (from peft->-r requirements_npu.txt (line 2)) Requirement already satisfied: xxhash in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.4.1) Requirement already satisfied: multiprocess in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (0.70.15) Requirement already satisfied: fsspec<=2024.6.1,>=2023.1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from fsspec[http]<=2024.6.1,>=2023.1.0->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2023.10.0) Requirement already satisfied: aiohttp in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.9.3) Processing data/software/huggingface_hub-0.26.2-py3-none-any.whl (from peft->-r requirements_npu.txt (line 2)) Requirement already satisfied: protobuf>=3.20 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tensorboardX->-r requirements_npu.txt (line 11)) (4.25.3) Requirement already satisfied: tokenizers<0.20,>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers>=4.40.1->-r requirements_npu.txt (line 13)) (0.19.1) Requirement already satisfied: hjson in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (3.1.0) Requirement already satisfied: ninja in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (1.11.1.1) Requirement already satisfied: nvidia-ml-py in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (12.560.30) Requirement already satisfied: py-cpuinfo in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (9.0.0) Requirement already satisfied: pydantic in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (1.10.15) Processing data/software/safetensors-0.4.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from peft->-r requirements_npu.txt (line 2)) Requirement already satisfied: torchvision in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from timm->-r requirements_npu.txt (line 16)) (0.16.0) Requirement already satisfied: Werkzeug>=3.0.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (3.0.1) Requirement already satisfied: Jinja2>=3.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (3.1.3) Processing data/software/itsdangerous-2.2.0-py3-none-any.whl (from flask->-r requirements_npu.txt (line 17)) Requirement already satisfied: click>=8.1.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (8.1.7) Processing data/software/blinker-1.8.2-py3-none-any.whl (from flask->-r requirements_npu.txt (line 17)) Requirement already satisfied: importlib-metadata>=3.6.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (7.0.1) Processing data/software/aniso8601-9.0.1-py2.py3-none-any.whl (from flask_restful->-r requirements_npu.txt (line 18)) Requirement already satisfied: six>=1.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask_restful->-r requirements_npu.txt (line 18)) (1.16.0) Requirement already satisfied: pytz in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask_restful->-r requirements_npu.txt (line 18)) (2024.1) Requirement already satisfied: aiosignal>=1.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.3.1) Requirement already satisfied: attrs>=17.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (23.2.0) Requirement already satisfied: frozenlist>=1.1.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.4.1) Requirement already satisfied: multidict<7.0,>=4.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (6.0.5) Requirement already satisfied: yarl<2.0,>=1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.9.4) Requirement already satisfied: async-timeout<5.0,>=4.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (4.0.3) Requirement already satisfied: typing-extensions>=3.7.4.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from huggingface-hub>=0.17.0->peft->-r requirements_npu.txt (line 2)) (4.10.0) Requirement already satisfied: zipp>=0.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from importlib-metadata>=3.6.0->flask->-r requirements_npu.txt (line 17)) (3.17.0) Requirement already satisfied: MarkupSafe>=2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from Jinja2>=3.1.2->flask->-r requirements_npu.txt (line 17)) (2.1.5) Requirement already satisfied: charset-normalizer<4,>=2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (3.3.2) Requirement already satisfied: idna<4,>=2.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (3.6) Requirement already satisfied: urllib3<3,>=1.21.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (1.26.18) Requirement already satisfied: certifi>=2017.4.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (2024.2.2) Requirement already satisfied: sympy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (1.4) Requirement already satisfied: networkx in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (3.1) Requirement already satisfied: tzdata>=2022.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2024.1) Requirement already satisfied: mpmath>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from sympy->torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (1.3.0) DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063 Installing collected packages: aniso8601, xlsxwriter, tqdm, safetensors, requests, pybind11, pyarrow, natsort, itsdangerous, expecttest, decord, blinker, tiktoken, huggingface-hub, flask, flask_restful, accelerate, datasets Attempting uninstall: tqdm Found existing installation: tqdm 4.66.2 Uninstalling tqdm-4.66.2: Successfully uninstalled tqdm-4.66.2 Attempting uninstall: safetensors Found existing installation: safetensors 0.4.2 Uninstalling safetensors-0.4.2: Successfully uninstalled safetensors-0.4.2 Attempting uninstall: requests Found existing installation: requests 2.31.0 Uninstalling requests-2.31.0: Successfully uninstalled requests-2.31.0 Attempting uninstall: pyarrow Found existing installation: pyarrow 15.0.0 Uninstalling pyarrow-15.0.0: Successfully uninstalled pyarrow-15.0.0 Attempting uninstall: huggingface-hub Found existing installation: huggingface-hub 0.20.3 Uninstalling huggingface-hub-0.20.3: Successfully uninstalled huggingface-hub-0.20.3 Attempting uninstall: accelerate Found existing installation: accelerate 0.25.0 Uninstalling accelerate-0.25.0: Successfully uninstalled accelerate-0.25.0 Attempting uninstall: datasets Found existing installation: datasets 2.16.0 Uninstalling datasets-2.16.0: Successfully uninstalled datasets-2.16.0 ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. tikit 1.8.2.240926 requires dicttoxml==1.7.4, which is not installed. tikit 1.8.2.240926 requires docopt==0.6.2, which is not installed. tikit 1.8.2.240926 requires future==0.18.2, which is not installed. tikit 1.8.2.240926 requires hdfs==2.6.0, which is not installed. tikit 1.8.2.240926 requires pure-sasl==0.6.2, which is not installed. tikit 1.8.2.240926 requires py4j==0.10.7, which is not installed. tikit 1.8.2.240926 requires PyHive[hive]==0.6.4, which is not installed. tikit 1.8.2.240926 requires pyjwt>=2.4.0, which is not installed. tikit 1.8.2.240926 requires requests-kerberos>=0.14.0, which is not installed. tikit 1.8.2.240926 requires sasl==0.3.1, which is not installed. tikit 1.8.2.240926 requires thrift==0.15.0, which is not installed. tikit 1.8.2.240926 requires thrift-sasl>=0.1.0, which is not installed. tikit 1.8.2.240926 requires certifi==2021.10.8, but you have certifi 2024.2.2 which is incompatible. tikit 1.8.2.240926 requires cos-python-sdk-v5==1.9.29, but you have cos-python-sdk-v5 1.9.26 which is incompatible. tikit 1.8.2.240926 requires idna==3.3, but you have idna 3.6 which is incompatible. tikit 1.8.2.240926 requires prettytable==2.5.0, but you have prettytable 3.11.0 which is incompatible. tikit 1.8.2.240926 requires urllib3==1.26.7, but you have urllib3 1.26.18 which is incompatible. tikit 1.8.2.240926 requires wcwidth==0.2.5, but you have wcwidth 0.2.13 which is incompatible. Successfully installed accelerate-0.34.2 aniso8601-9.0.1 blinker-1.8.2 datasets-3.0.0 decord-0.6.0 expecttest-0.2.1 flask-3.0.3 flask_restful-0.3.10 huggingface-hub-0.26.2 itsdangerous-2.2.0 natsort-8.4.0 pyarrow-17.0.0 pybind11-2.13.6 requests-2.32.3 safetensors-0.4.5 tiktoken-0.7.0 tqdm-4.67.1 xlsxwriter-3.2.0 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv ++ return 0 + MEGATRON_DIR=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0/ + MINDSPEED_DIR=/local_disk/cognitron_vl//third_party/MindSpeed_core_r0.6.0/ + MODELLINK_DIR=/local_disk/cognitron_vl//third_party/ModelLink/ + pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0/ Looking in links: /data/software/ Obtaining file://local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0 Installing build dependencies: started Installing build dependencies: finished with status 'done' Checking if build backend supports build_editable: started Checking if build backend supports build_editable: finished with status 'done' Getting requirements to build editable: started Getting requirements to build editable: finished with status 'done' Installing backend dependencies: started Installing backend dependencies: finished with status 'done' Preparing editable metadata (pyproject.toml): started Preparing editable metadata (pyproject.toml): finished with status 'done' Building wheels for collected packages: megatron_core Building editable for megatron_core (pyproject.toml): started Building editable for megatron_core (pyproject.toml): finished with status 'done' Created wheel for megatron_core: filename=megatron_core-0.6.0-0.editable-cp38-cp38-linux_x86_64.whl size=8791 sha256=aa9f6bbbd19fa94cd58cbe4b8f2a2d826dd8073675a29497a20e32c8e06c1d50 Stored in directory: /tmp/pip-ephem-wheel-cache-2_mx6hhf/wheels/54/9c/d1/d2015aa0c34e791e64d65d19395e5a9a5528f0c63fd519b9ff Successfully built megatron_core DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063 Installing collected packages: megatron_core Successfully installed megatron_core-0.6.0 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv + pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/MindSpeed_core_r0.6.0/ Looking in links: /data/software/ Obtaining file://local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0 Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' WARNING: Error parsing requirements for tokenizers: [Errno 2] No such file or directory: '/root/miniconda3/envs/py38/lib/python3.8/site-packages/tokenizers-0.19.1.dist-info/METADATA' WARNING: Error parsing requirements for transformers: [Errno 2] No such file or directory: '/root/miniconda3/envs/py38/lib/python3.8/site-packages/transformers-4.40.1.dist-info/METADATA' DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063 Installing collected packages: mindspeed Running setup.py develop for mindspeed Successfully installed mindspeed-0.6.0 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv + pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/ModelLink/ Looking in links: /data/software/ Obtaining file://local_disk/cognitron_vl/third_party/ModelLink Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Requirement already satisfied: numpy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.24.4) Processing data/software/transformers-4.43.2-py3-none-any.whl (from modellink==0.0.1) Processing data/software/transformers-stream-generator-0.0.5.tar.gz (from modellink==0.0.1) Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Requirement already satisfied: sympy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.4) Requirement already satisfied: decorator in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (5.1.1) Requirement already satisfied: scipy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.10.1) Requirement already satisfied: sentencepiece in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.2.0) Requirement already satisfied: einops in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.7.0) Requirement already satisfied: datasets in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (3.0.0) Requirement already satisfied: pybind11 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (2.13.6) Requirement already satisfied: accelerate in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.34.2) Requirement already satisfied: six in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.16.0) Requirement already satisfied: protobuf in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (4.25.3) Processing data/software/peft-0.7.1-py3-none-any.whl (from modellink==0.0.1) Requirement already satisfied: tiktoken in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.7.0) Requirement already satisfied: packaging>=20.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (23.2) Requirement already satisfied: psutil in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (5.9.8) Requirement already satisfied: pyyaml in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (5.4.1) Requirement already satisfied: torch>=1.13.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (2.1.0+cpu) Requirement already satisfied: tqdm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (4.67.1) Requirement already satisfied: safetensors in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (0.4.5) Requirement already satisfied: huggingface-hub>=0.17.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (0.26.2) Requirement already satisfied: filelock in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (3.13.1) Requirement already satisfied: regex!=2019.12.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (2023.12.25) Requirement already satisfied: requests in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (2.32.3) Processing data/software/tokenizers-0.19.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from transformers==4.43.2->modellink==0.0.1) Requirement already satisfied: pyarrow>=15.0.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (17.0.0) Requirement already satisfied: dill<0.3.9,>=0.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (0.3.7) Requirement already satisfied: pandas in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (2.0.3) Requirement already satisfied: xxhash in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (3.4.1) Requirement already satisfied: multiprocess in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (0.70.15) Requirement already satisfied: fsspec<=2024.6.1,>=2023.1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from fsspec[http]<=2024.6.1,>=2023.1.0->datasets->modellink==0.0.1) (2023.10.0) Requirement already satisfied: aiohttp in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (3.9.3) Requirement already satisfied: mpmath>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from sympy->modellink==0.0.1) (1.3.0) Requirement already satisfied: aiosignal>=1.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.3.1) Requirement already satisfied: attrs>=17.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (23.2.0) Requirement already satisfied: frozenlist>=1.1.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.4.1) Requirement already satisfied: multidict<7.0,>=4.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (6.0.5) Requirement already satisfied: yarl<2.0,>=1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.9.4) Requirement already satisfied: async-timeout<5.0,>=4.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (4.0.3) Requirement already satisfied: typing-extensions>=3.7.4.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from huggingface-hub>=0.17.0->peft==0.7.1->modellink==0.0.1) (4.10.0) Requirement already satisfied: charset-normalizer<4,>=2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (3.3.2) Requirement already satisfied: idna<4,>=2.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (3.6) Requirement already satisfied: urllib3<3,>=1.21.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (1.26.18) Requirement already satisfied: certifi>=2017.4.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (2024.2.2) Requirement already satisfied: networkx in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (3.1) Requirement already satisfied: jinja2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (3.1.3) Requirement already satisfied: python-dateutil>=2.8.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2.8.2) Requirement already satisfied: pytz>=2020.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2024.1) Requirement already satisfied: tzdata>=2022.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2024.1) Requirement already satisfied: MarkupSafe>=2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from jinja2->torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (2.1.5) Building wheels for collected packages: transformers_stream_generator Building wheel for transformers_stream_generator (setup.py): started Building wheel for transformers_stream_generator (setup.py): finished with status 'done' Created wheel for transformers_stream_generator: filename=transformers_stream_generator-0.0.5-py3-none-any.whl size=12425 sha256=7b79cc70e3e25aee266f004baeb02f49a26365cbf13fc2e075375f124a308963 Stored in directory: /root/.cache/pip/wheels/56/8c/42/5381d9c36bc85f28982f4cf8f98dc44d37a6d6c04897a5cb7c Successfully built transformers_stream_generator DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063 Installing collected packages: tokenizers, transformers, transformers_stream_generator, peft, modellink Attempting uninstall: tokenizers Found existing installation: tokenizers 0.20.3 Uninstalling tokenizers-0.20.3: Successfully uninstalled tokenizers-0.20.3 Attempting uninstall: transformers Found existing installation: transformers 4.46.3 Uninstalling transformers-4.46.3: Successfully uninstalled transformers-4.46.3 Attempting uninstall: peft Found existing installation: peft 0.7.0 Uninstalling peft-0.7.0: Successfully uninstalled peft-0.7.0 Running setup.py develop for modellink Successfully installed modellink-0.0.1 peft-0.7.1 tokenizers-0.19.1 transformers-4.43.2 transformers_stream_generator-0.0.5 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv + export PYTHONPATH=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0//:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe: + PYTHONPATH=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0//:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe: + GPUS_PER_NODE=16 + NNODES=32 + NODE_RANK=10 + MASTER_PORT=34567 + export CUDA_DEVICE_MAX_CONNECTIONS=1 + CUDA_DEVICE_MAX_CONNECTIONS=1 + export PYTORCH_NPU_ALLOC_CONF=expandable_segments:True + PYTORCH_NPU_ALLOC_CONF=expandable_segments:True + VISION_SEQ_LENGTH=1025 + IMAGE_TOKEN_LENGTH=256 + IMAGE_SIZE=448 + VISION_MODEL_TYPE=intern_300m + TP=8 + PP=1 + CP=8 + CP_ALGO=megatron_cp_algo + CP_MASK=causal + DISTRIBUTED_ARGS=' --nproc_per_node 16 --nnodes 32 --node_rank 10 --master_addr train-1198772881325351168-93vlj4s2getc-master-0.train-100034032793.svc.cluster.local --master_port 34567 ' + GPT_ARGS=' --use-mcore-models --tensor-model-parallel-size 8 --pipeline-model-parallel-size 1 --context-parallel-size 8 --context-parallel-algo megatron_cp_algo --cp-attention-mask-type causal --use-cp-send-recv-overlap --no-create-attention-mask-in-dataloader --sparse-mode 4 --sequence-parallel --recompute-method block --recompute-granularity full --recompute-num-layers 48 --num-layers 48 --hidden-size 5120 --ffn-hidden-size 13824 --num-attention-heads 40 --group-query-attention --num-query-groups 8 --tokenizer-type PretrainedFromHF --tokenizer-name-or-path /data_4/models/Qwen/Qwen2.5-14B-Instruct/ --seq-length 1048576 --max-position-embeddings 1048576 --micro-batch-size 1 --global-batch-size 8 --make-vocab-size-divisible-by 1 --padded-vocab-size 152064 --rotary-base 1000000.0 --lr 5.00e-6 --train-iters 500 --lr-decay-style cosine --untie-embeddings-and-output-weights --disable-bias-linear --attention-dropout 0.0 --init-method-std 0.01 --hidden-dropout 0.0 --position-embedding-type rope --normalization RMSNorm --use-fused-rmsnorm --norm-epsilon 1e-6 --swiglu --use-flash-attn --use-fused-rotary-pos-emb --use-rotary-position-embeddings --use-fused-swiglu --use-mc2 --no-masked-softmax-fusion --attention-softmax-in-fp32 --min-lr 1.00e-7 --weight-decay 0.0 --lr-warmup-fraction 0.03 --clip-grad 1.0 --adam-beta1 0.9 --adam-beta2 0.999 --add-qkv-bias --initial-loss-scale 4096 --no-gradient-accumulation-fusion --use-distributed-optimizer --bf16 --overlap-grad-reduce --finetune --vision-model-freeze --vision-model-type intern_300m --vision-downsample-ratio 0.5 --vision-projector-type mlp --vision-projector-pre-norm --vision-process-type dynamic --vision-normalize-type imagenet --vision-seq-length 1025 --image-token-length 256 --image-size 448 --prompt-format qwen2 --is-instruction-dataset --max-num-image 4096 --max-fps 1 --add-class-token --min-patch-grid 1 --max-patch-grid 12 --logit-mask --cross-dataset-joint ' + DATA_ARGS=' --data-path /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml --split 100,0,0 --data-seq-length 1048576 --num-workers 8 ' + CKPT_ARGS=' --load /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/ --vit-load / --no-load-optim --no-load-rng --seed 42424242 --save /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743// ' + OUTPUT_ARGS=' --log-interval 1 --save-interval 20 --eval-interval 20 --eval-iters 0 --log-throughput --distributed-timeout-minutes 120 ' + torchrun --nproc_per_node 16 --nnodes 32 --node_rank 10 --master_addr train-1198772881325351168-93vlj4s2getc-master-0.train-100034032793.svc.cluster.local --master_port 34567 /local_disk/cognitron_vl//lcvlm_modellink/pretrain_lcvlm.py --use-mcore-models --tensor-model-parallel-size 8 --pipeline-model-parallel-size 1 --context-parallel-size 8 --context-parallel-algo megatron_cp_algo --cp-attention-mask-type causal --use-cp-send-recv-overlap --no-create-attention-mask-in-dataloader --sparse-mode 4 --sequence-parallel --recompute-method block --recompute-granularity full --recompute-num-layers 48 --num-layers 48 --hidden-size 5120 --ffn-hidden-size 13824 --num-attention-heads 40 --group-query-attention --num-query-groups 8 --tokenizer-type PretrainedFromHF --tokenizer-name-or-path /data_4/models/Qwen/Qwen2.5-14B-Instruct/ --seq-length 1048576 --max-position-embeddings 1048576 --micro-batch-size 1 --global-batch-size 8 --make-vocab-size-divisible-by 1 --padded-vocab-size 152064 --rotary-base 1000000.0 --lr 5.00e-6 --train-iters 500 --lr-decay-style cosine --untie-embeddings-and-output-weights --disable-bias-linear --attention-dropout 0.0 --init-method-std 0.01 --hidden-dropout 0.0 --position-embedding-type rope --normalization RMSNorm --use-fused-rmsnorm --norm-epsilon 1e-6 --swiglu --use-flash-attn --use-fused-rotary-pos-emb --use-rotary-position-embeddings --use-fused-swiglu --use-mc2 --no-masked-softmax-fusion --attention-softmax-in-fp32 --min-lr 1.00e-7 --weight-decay 0.0 --lr-warmup-fraction 0.03 --clip-grad 1.0 --adam-beta1 0.9 --adam-beta2 0.999 --add-qkv-bias --initial-loss-scale 4096 --no-gradient-accumulation-fusion --use-distributed-optimizer --bf16 --overlap-grad-reduce --finetune --vision-model-freeze --vision-model-type intern_300m --vision-downsample-ratio 0.5 --vision-projector-type mlp --vision-projector-pre-norm --vision-process-type dynamic --vision-normalize-type imagenet --vision-seq-length 1025 --image-token-length 256 --image-size 448 --prompt-format qwen2 --is-instruction-dataset --max-num-image 4096 --max-fps 1 --add-class-token --min-patch-grid 1 --max-patch-grid 12 --logit-mask --cross-dataset-joint --data-path /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml --split 100,0,0 --data-seq-length 1048576 --num-workers 8 --load /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/ --vit-load / --no-load-optim --no-load-rng --seed 42424242 --save /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743// --log-interval 1 --save-interval 20 --eval-interval 20 --eval-iters 0 --log-throughput --distributed-timeout-minutes 120 --distributed-backend nccl [2024-11-28 15:50:28,011] torch.distributed.run: [WARNING] [2024-11-28 15:50:28,011] torch.distributed.run: [WARNING] ***************************************** [2024-11-28 15:50:28,011] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. [2024-11-28 15:50:28,011] torch.distributed.run: [WARNING] ***************************************** Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Creating extension directory /root/.cache/torch_extensions/py38_cpu/adaptive_cp... Emitting ninja build file /root/.cache/torch_extensions/py38_cpu/adaptive_cp/build.ninja... Building extension module adaptive_cp... Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root... [1/2] c++ -MMD -MF adaptive_cp.o.d -DTORCH_EXTENSION_NAME=adaptive_cp -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_gcc\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1011\" -I/usr/local/Ascend/ascend-toolkit/latest/include -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/include -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/third_party -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/acl -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/inc -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/TH -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/THC -isystem /root/miniconda3/envs/py38/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=0 -fPIC -std=c++17 -fstack-protector-all -Wl,-z,relro,-z,now,-z,noexecstack -fPIC -pie -Wl,--disable-new-dtags,--rpath -s -O2 -c local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/ops/csrc/algorithm/adaptive_cp/adaptive_cp.cpp -o adaptive_cp.o [2/2] c++ adaptive_cp.o -shared -L/usr/local/Ascend/ascend-toolkit/latest/lib64 -lascendcl -L/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/lib -ltorch_npu -L/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o adaptive_cp.so Loading extension module adaptive_cp... Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... Loading extension module adaptive_cp... Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") Loading extension module adaptive_cp... local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32 warnings.warn("failed to generate the npu_matmul_add_fp32") /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( /root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source? warn( > compiling dataset index builder ... make: Entering directory 'local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0/megatron/core/datasets' make: Nothing to be done for 'default'. make: Leaving directory 'local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0/megatron/core/datasets' >>> done with dataset index builder. Compilation time: 0.245 seconds vision_projector_recompute False vision_projector_recompute False vision_projector_recompute Falsevision_projector_recompute Falsevision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute False vision_projector_recompute Falsevision_projector_recompute False vision_projector_recompute False vision_model_freeze => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. vision_model_freeze=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. vision_model_freeze=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. vision_model_freeze => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. vision_model_freeze=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. vision_model_freeze => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. vision_model_freeze=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.vision_model_freeze => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.vision_model_freeze=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.vision_model_freeze => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.vision_model_freeze=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) vision_model_freeze => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) vision_model_freeze vision_model_freeze => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) vision_model_freeze => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. vision_model_freeze => set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False. => set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False. => set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() )=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False. => set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False. model GPTVLModel( (external_feature_model): MegatronVisionModel( (vit): InternViTModel( (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14)) (position_embeddings): Embedding(1025, 1024) (decoder): TransformerBlock( (layers): ModuleList( (0-23): 24 x InternViTTransformerLayer( (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() ) (self_attn_bda): IdentityFuncOp() (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True) (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) (mlp_bda): IdentityFuncOp() ) ) ) ) (vision_projection): MultimodalProjector( (encoder): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True) ) (embedding): LanguageModelEmbedding( (word_embeddings): VocabParallelEmbedding() (embedding_dropout): Dropout(p=0.0, inplace=False) ) (rotary_pos_emb): RotaryEmbedding() (decoder): TransformerBlock( (layers): ModuleList( (0-47): 48 x TransformerLayer( (input_layernorm): RMSNorm() (self_attention): SelfAttention( (core_attention): DotProductAttention( (scale_mask_softmax): FusedScaleMaskSoftmax() (attention_dropout): Dropout(p=0.0, inplace=False) ) (linear_proj): RowParallelLinear() (linear_qkv): ColumnParallelLinear() (q_layernorm): IdentityOp() (k_layernorm): IdentityOp() ) (pre_cross_attn_layernorm): IdentityOp() (cross_attention): IdentityOp() (cross_attn_bda): IdentityFuncOp() (pre_mlp_layernorm): RMSNorm() (mlp): MLP( (linear_fc1): ColumnParallelLinear() (linear_fc2): RowParallelLinear() ) ) ) (final_layernorm): RMSNorm() ) (output_layer): ColumnParallelLinear() ) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False) _get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False) _get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False) _load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration _load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration _load_base_checkpoint iteration1000 _load_base_checkpoint iteration_load_base_checkpoint iteration1000_load_base_checkpoint iteration 10001000 1000 1000 1000 _load_base_checkpoint release 10001000 1000 1000 1000_load_base_checkpoint release1000 _load_base_checkpoint release False _load_base_checkpoint release _load_base_checkpoint releaseFalse1000_load_base_checkpoint release1000 _load_base_checkpoint release_load_base_checkpoint release_load_base_checkpoint release _load_base_checkpoint release FalseFalse _load_base_checkpoint release False 1000 _load_base_checkpoint release False _load_base_checkpoint release_load_base_checkpoint release FalseFalseFalse False_load_base_checkpoint release FalseFalse FalseFalse _load_base_checkpoint release False False _load_base_checkpoint_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_07/model_optim_rng.pt/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_01/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_00/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_05/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_04/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_05/model_optim_rng.pt _load_base_checkpoint_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_06/model_optim_rng.pt/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_02/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_07/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_06/model_optim_rng.pt _load_base_checkpoint_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_00/model_optim_rng.pt/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_04/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_02/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_03/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_03/model_optim_rng.pt _load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_01/model_optim_rng.pt load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True load_checkpoint iteration 0 load_checkpoint release False strict True > rank 164 does not create GPT datasets ... > rank 165 does not create GPT datasets ... > rank 169 does not create GPT datasets ... > rank 161 does not create GPT datasets ... > rank 167 does not create GPT datasets ... > rank 172 does not create GPT datasets ... > rank 160 is creating GPT datasets ... > rank 170 does not create GPT datasets ... > rank 173 does not create GPT datasets ...> rank 168 is creating GPT datasets ...> rank 175 does not create GPT datasets ... > rank 162 does not create GPT datasets ... > rank 174 does not create GPT datasets ... > rank 166 does not create GPT datasets ... > rank 163 does not create GPT datasets ... > rank 171 does not create GPT datasets ... target_ratios [(1, 1), (1, 2), (2, 1), (3, 1), (1, 3), (2, 2), (4, 1), (1, 4), (5, 1), (1, 5), (1, 6), (6, 1), (3, 2), (2, 3), (7, 1), (1, 7), (4, 2), (2, 4), (1, 8), (8, 1), (1, 9), (3, 3), (9, 1), (2, 5), (5, 2), (10, 1), (1, 10), (11, 1), (1, 11), (12, 1), (3, 4), (4, 3), (1, 12), (6, 2), (2, 6)] target_ratios [(1, 1), (1, 2), (2, 1), (3, 1), (1, 3), (2, 2), (4, 1), (1, 4), (5, 1), (1, 5), (1, 6), (6, 1), (3, 2), (2, 3), (7, 1), (1, 7), (4, 2), (2, 4), (1, 8), (8, 1), (1, 9), (3, 3), (9, 1), (2, 5), (5, 2), (10, 1), (1, 10), (11, 1), (1, 11), (12, 1), (3, 4), (4, 3), (1, 12), (6, 2), (2, 6)] possible_resolutions [[448, 448], [448, 896], [896, 448], [1344, 448], [448, 1344], [896, 896], [1792, 448], [448, 1792], [2240, 448], [448, 2240], [448, 2688], [2688, 448], [1344, 896], [896, 1344], [3136, 448], [448, 3136], [1792, 896], [896, 1792], [448, 3584], [3584, 448], [448, 4032], [1344, 1344], [4032, 448], [896, 2240], [2240, 896], [4480, 448], [448, 4480], [4928, 448], [448, 4928], [5376, 448], [1344, 1792], [1792, 1344], [448, 5376], [2688, 896], [896, 2688]] possible_resolutions [[448, 448], [448, 896], [896, 448], [1344, 448], [448, 1344], [896, 896], [1792, 448], [448, 1792], [2240, 448], [448, 2240], [448, 2688], [2688, 448], [1344, 896], [896, 1344], [3136, 448], [448, 3136], [1792, 896], [896, 1792], [448, 3584], [3584, 448], [448, 4032], [1344, 1344], [4032, 448], [896, 2240], [2240, 896], [4480, 448], [448, 4480], [4928, 448], [448, 4928], [5376, 448], [1344, 1792], [1792, 1344], [448, 5376], [2688, 896], [896, 2688]] [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea2e6cfc0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure processed_samples 100 unjoint_samples 100 joint_samples 0 [186086, 155645] processed_samples 100 unjoint_samples 100 joint_samples 0 [186086, 155645] processed_samples 100 unjoint_samples 100 joint_samples 0 [211662, 188985] processed_samples 100 unjoint_samples 100 joint_samples 0 [211662, 188985] processed_samples 100 unjoint_samples 100 joint_samples 0 [126884, 126011] processed_samples 100 unjoint_samples 100 joint_samples 0 [126884, 126011] processed_samples 100 unjoint_samples 100 joint_samples 0 [142367, 158447] processed_samples 100 unjoint_samples 100 joint_samples 0 [142367, 158447] processed_samples 100 unjoint_samples 100 joint_samples 0 [152264, 136690] processed_samples 100 unjoint_samples 100 joint_samples 0 [152264, 136690] processed_samples 100 unjoint_samples 100 joint_samples 0 [151012, 151523] processed_samples 100 unjoint_samples 100 joint_samples 0 [151012, 151523] processed_samples 100 unjoint_samples 100 joint_samples 0 [128892, 130133] processed_samples 100 unjoint_samples 100 joint_samples 0 [128892, 130133] processed_samples 100 unjoint_samples 100 joint_samples 0 [142191, 137405] processed_samples 100 unjoint_samples 100 joint_samples 0 [142191, 137405] [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure processed_samples 200 unjoint_samples 200 joint_samples 0 [328847, 343359] processed_samples 200 unjoint_samples 200 joint_samples 0 [328847, 343359] processed_samples 200 unjoint_samples 200 joint_samples 0 [303024, 296898] processed_samples 200 unjoint_samples 200 joint_samples 0 [303024, 296898] processed_samples 200 unjoint_samples 200 joint_samples 0 [310969, 322824] processed_samples 200 unjoint_samples 200 joint_samples 0 [310969, 322824] processed_samples 200 unjoint_samples 200 joint_samples 0 [320774, 310208] processed_samples 200 unjoint_samples 200 joint_samples 0 [284798, 287403] processed_samples 200 unjoint_samples 200 joint_samples 0 [317589, 340911] processed_samples 200 unjoint_samples 200 joint_samples 0 [320774, 310208] processed_samples 200 unjoint_samples 200 joint_samples 0 [319065, 319150] processed_samples 200 unjoint_samples 200 joint_samples 0 [284798, 287403] processed_samples 200 unjoint_samples 200 joint_samples 0 [317589, 340911] processed_samples 200 unjoint_samples 200 joint_samples 0 [319065, 319150] processed_samples 200 unjoint_samples 200 joint_samples 0 [250143, 251280] processed_samples 200 unjoint_samples 200 joint_samples 0 [250143, 251280] [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 300 unjoint_samples 300 joint_samples 0 [451226, 451791] processed_samples 300 unjoint_samples 300 joint_samples 0 [451226, 451791] processed_samples 300 unjoint_samples 300 joint_samples 0 [470625, 463973] processed_samples 300 unjoint_samples 300 joint_samples 0 [475161, 475142] processed_samples 300 unjoint_samples 300 joint_samples 0 [470625, 463973] processed_samples 300 unjoint_samples 300 joint_samples 0 [475161, 475142] processed_samples 300 unjoint_samples 300 joint_samples 0 [539320, 549278] processed_samples 300 unjoint_samples 300 joint_samples 0 [539320, 549278] processed_samples 300 unjoint_samples 300 joint_samples 0 [471198, 469420] processed_samples 300 unjoint_samples 300 joint_samples 0 [471198, 469420] processed_samples 300 unjoint_samples 300 joint_samples 0 [374724, 375819] processed_samples 300 unjoint_samples 300 joint_samples 0 [374724, 375819] [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure processed_samples 300 unjoint_samples 300 joint_samples 0 [449383, 446698] processed_samples 300 unjoint_samples 300 joint_samples 0 [449383, 446698] processed_samples 300 unjoint_samples 300 joint_samples 0 [472420, 470497] processed_samples 300 unjoint_samples 300 joint_samples 0 [472420, 470497] [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure processed_samples 400 unjoint_samples 400 joint_samples 0 [594638, 568912] processed_samples 400 unjoint_samples 400 joint_samples 0 [594638, 568912] processed_samples 400 unjoint_samples 400 joint_samples 0 [611620, 605888] processed_samples 400 unjoint_samples 400 joint_samples 0 [611620, 605888] processed_samples 400 unjoint_samples 400 joint_samples 0 [492414, 492782] processed_samples 400 unjoint_samples 400 joint_samples 0 [624310, 622805] processed_samples 400 unjoint_samples 400 joint_samples 0 [708562, 709703] processed_samples 400 unjoint_samples 400 joint_samples 0 [492414, 492782] processed_samples 400 unjoint_samples 400 joint_samples 0 [708562, 709703] processed_samples 400 unjoint_samples 400 joint_samples 0 [624310, 622805] [h264 @ 0x55e6d16cf440] Invalid NAL unit size (1113487007 > 108878). [h264 @ 0x55e6d16cf440] Error splitting the input into NAL units. [h264 @ 0x561ea4251980] Invalid NAL unit size (1113487007 > 108878). [h264 @ 0x561ea4251980] Error splitting the input into NAL units. processed_samples 401 unjoint_samples 400 joint_samples 0 [697811, 697683] processed_samples 401 unjoint_samples 400 joint_samples 0 [697811, 697683] processed_samples 400 unjoint_samples 400 joint_samples 0 [638103, 639615] processed_samples 400 unjoint_samples 400 joint_samples 0 [638103, 639615] processed_samples 400 unjoint_samples 400 joint_samples 0 [584342, 581075] processed_samples 400 unjoint_samples 400 joint_samples 0 [584342, 581075] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure processed_samples 500 unjoint_samples 500 joint_samples 0 [839195, 838946] processed_samples 501 unjoint_samples 500 joint_samples 0 [839133, 843215] processed_samples 500 unjoint_samples 500 joint_samples 0 [839195, 838946] processed_samples 501 unjoint_samples 500 joint_samples 0 [839133, 843215] processed_samples 500 unjoint_samples 500 joint_samples 0 [795485, 796170] processed_samples 500 unjoint_samples 500 joint_samples 0 [795485, 796170] processed_samples 500 unjoint_samples 500 joint_samples 0 [701466, 703343] processed_samples 500 unjoint_samples 500 joint_samples 0 [701466, 703343] processed_samples 500 unjoint_samples 500 joint_samples 0 [718962, 717422] processed_samples 500 unjoint_samples 500 joint_samples 0 [640068, 641089] processed_samples 500 unjoint_samples 500 joint_samples 0 [640068, 641089] processed_samples 500 unjoint_samples 500 joint_samples 0 [718962, 717422] processed_samples 500 unjoint_samples 500 joint_samples 0 [763123, 762026] processed_samples 500 unjoint_samples 500 joint_samples 0 [763123, 762026] processed_samples 500 unjoint_samples 500 joint_samples 0 [778057, 779272] processed_samples 500 unjoint_samples 500 joint_samples 0 [778057, 779272] [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6c03b2000] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure processed_samples 600 unjoint_samples 600 joint_samples 0 [917982, 874398] processed_samples 600 unjoint_samples 600 joint_samples 0 [917982, 874398] processed_samples 600 unjoint_samples 600 joint_samples 0 [760499, 762276] processed_samples 600 unjoint_samples 600 joint_samples 0 [941132, 943417] processed_samples 600 unjoint_samples 600 joint_samples 0 [833932, 834377] processed_samples 600 unjoint_samples 600 joint_samples 0 [760499, 762276] processed_samples 600 unjoint_samples 600 joint_samples 0 [941132, 943417] processed_samples 600 unjoint_samples 600 joint_samples 0 [833932, 834377] processed_samples 600 unjoint_samples 600 joint_samples 0 [918037, 927245] processed_samples 600 unjoint_samples 600 joint_samples 0 [936661, 973156] processed_samples 600 unjoint_samples 600 joint_samples 0 [936661, 973156] processed_samples 600 unjoint_samples 600 joint_samples 0 [918037, 927245] processed_samples 600 unjoint_samples 600 joint_samples 0 [969016, 968838] processed_samples 600 unjoint_samples 600 joint_samples 0 [969016, 968838] processed_samples 601 unjoint_samples 600 joint_samples 0 [983580, 985069] processed_samples 601 unjoint_samples 600 joint_samples 0 [983580, 985069] [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure ................................................................................................[h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6c03b2000] mmco: unref short failure [h264 @ 0x55e6c03b2000] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure processed_samples 700 unjoint_samples 700 joint_samples 1 [1046636, 98311] processed_samples 700 unjoint_samples 700 joint_samples 1 [1046636, 98311] [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 700 unjoint_samples 700 joint_samples 1 [1045112, 115548] processed_samples 700 unjoint_samples 700 joint_samples 1 [1045112, 115548] processed_samples 700 unjoint_samples 700 joint_samples 1 [237768, 1046424] processed_samples 700 unjoint_samples 700 joint_samples 1 [237768, 1046424] processed_samples 701 unjoint_samples 700 joint_samples 1 [1029709, 281082] processed_samples 701 unjoint_samples 700 joint_samples 1 [1029709, 281082] processed_samples 700 unjoint_samples 700 joint_samples 0 [1037724, 1038409] processed_samples 700 unjoint_samples 700 joint_samples 0 [1037724, 1038409] processed_samples 700 unjoint_samples 700 joint_samples 0 [1042453, 1046328] processed_samples 700 unjoint_samples 700 joint_samples 0 [1042453, 1046328] processed_samples 700 unjoint_samples 700 joint_samples 0 [961347, 960047] processed_samples 700 unjoint_samples 700 joint_samples 0 [961347, 960047] processed_samples 700 unjoint_samples 700 joint_samples 0 [877767, 877825] processed_samples 700 unjoint_samples 700 joint_samples 0 [877767, 877825] [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure processed_samples 800 unjoint_samples 800 joint_samples 1 [193876, 1038899] processed_samples 800 unjoint_samples 800 joint_samples 1 [193876, 1038899] processed_samples 800 unjoint_samples 800 joint_samples 1 [605693, 1046424] [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure processed_samples 800 unjoint_samples 800 joint_samples 1 [1046636, 417130] processed_samples 800 unjoint_samples 800 joint_samples 1 [605693, 1046424] processed_samples 800 unjoint_samples 800 joint_samples 1 [242608, 1046323] processed_samples 800 unjoint_samples 800 joint_samples 1 [1046636, 417130] processed_samples 800 unjoint_samples 800 joint_samples 1 [1048171, 275215] processed_samples 800 unjoint_samples 800 joint_samples 1 [242608, 1046323] processed_samples 800 unjoint_samples 800 joint_samples 1 [1048171, 275215] processed_samples 800 unjoint_samples 800 joint_samples 1 [71757, 1024287] processed_samples 800 unjoint_samples 800 joint_samples 1 [71757, 1024287] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure processed_samples 800 unjoint_samples 800 joint_samples 1 [1045112, 491165] processed_samples 800 unjoint_samples 800 joint_samples 1 [1045112, 491165] processed_samples 801 unjoint_samples 800 joint_samples 1 [1029709, 558683] processed_samples 801 unjoint_samples 800 joint_samples 1 [1029709, 558683] [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure processed_samples 900 unjoint_samples 900 joint_samples 1 [609480, 1046323] processed_samples 900 unjoint_samples 900 joint_samples 1 [609480, 1046323] processed_samples 900 unjoint_samples 900 joint_samples 1 [677927, 1038899] processed_samples 900 unjoint_samples 900 joint_samples 1 [677927, 1038899] processed_samples 900 unjoint_samples 900 joint_samples 1 [1046636, 754475] processed_samples 900 unjoint_samples 900 joint_samples 1 [1046636, 754475] [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure processed_samples 900 unjoint_samples 900 joint_samples 1 [963130, 1046424] processed_samples 900 unjoint_samples 900 joint_samples 1 [1048171, 543630] processed_samples 900 unjoint_samples 900 joint_samples 1 [1048171, 543630] processed_samples 900 unjoint_samples 900 joint_samples 1 [963130, 1046424] processed_samples 900 unjoint_samples 900 joint_samples 1 [426950, 1024287] processed_samples 900 unjoint_samples 900 joint_samples 1 [426950, 1024287] processed_samples 901 unjoint_samples 900 joint_samples 1 [1029709, 797871] processed_samples 901 unjoint_samples 900 joint_samples 1 [1029709, 797871] processed_samples 900 unjoint_samples 900 joint_samples 1 [1045112, 765796] processed_samples 900 unjoint_samples 900 joint_samples 1 [1045112, 765796] [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6c03b2000] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6c03b2000] mmco: unref short failure [h264 @ 0x55e6c03b2000] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure processed_samples 1000 unjoint_samples 1000 joint_samples 2 [129078, 995495] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [129078, 995495] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [62769, 1039636] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [62769, 1039636] processed_samples 1001 unjoint_samples 1000 joint_samples 2 [1047971, 43561] processed_samples 1001 unjoint_samples 1000 joint_samples 2 [1047971, 43561] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [1048171, 882895] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [1048171, 882895] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [983577, 1046323] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [983577, 1046323] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [274999, 1046424] processed_samples 1000 unjoint_samples 1000 joint_samples 2 [274999, 1046424] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [1041938, 1042472] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [747685, 1024287] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [1041938, 1042472] processed_samples 1000 unjoint_samples 1000 joint_samples 1 [747685, 1024287] [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure processed_samples 1100 unjoint_samples 1100 joint_samples 2 [477835, 995495] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [477835, 995495] [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1039174, 27716] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1039174, 27716] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [87044, 1046804] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [87044, 1046804] processed_samples 1101 unjoint_samples 1100 joint_samples 2 [1047971, 382401] processed_samples 1101 unjoint_samples 1100 joint_samples 2 [1047971, 382401] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [615260, 1046424] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [615260, 1046424] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1045925, 272090] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1045925, 272090] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1047339, 323892] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1047339, 323892] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [411468, 1039636] processed_samples 1100 unjoint_samples 1100 joint_samples 2 [411468, 1039636] [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1047339, 606227] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1047339, 606227] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [376432, 1046804] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1045925, 571096] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1039174, 349326] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [376432, 1046804] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1045925, 571096] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1039174, 349326] [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure processed_samples 1200 unjoint_samples 1200 joint_samples 2 [684700, 1039636] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [684700, 1039636] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [952361, 1046424] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [952361, 1046424] processed_samples 1201 unjoint_samples 1200 joint_samples 2 [1047971, 630668] processed_samples 1201 unjoint_samples 1200 joint_samples 2 [1047971, 630668] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [707362, 995495] processed_samples 1200 unjoint_samples 1200 joint_samples 2 [707362, 995495] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure processed_samples 1300 unjoint_samples 1300 joint_samples 3 [1026091, 303089] processed_samples 1300 unjoint_samples 1300 joint_samples 3 [1026091, 303089] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [982301, 995495] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [982301, 995495] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1045925, 881459] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1045925, 881459] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1039174, 728413] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1039174, 728413] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [743693, 1046804] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1047339, 992559] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [743693, 1046804] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1047339, 992559] [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure processed_samples 1301 unjoint_samples 1300 joint_samples 2 [1047971, 908071] processed_samples 1301 unjoint_samples 1300 joint_samples 2 [1047971, 908071] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [999613, 1039636] processed_samples 1300 unjoint_samples 1300 joint_samples 2 [999613, 1039636] [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cec12f80] mmco: unref short failure [h264 @ 0x55e6cec12f80] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea3f9cc00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure processed_samples 1400 unjoint_samples 1400 joint_samples 3 [19043, 1037597]processed_samples 1400 unjoint_samples 1400 joint_samples 3 [19043, 1037597] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1047538, 142047] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1047538, 142047] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [83335, 1047343] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [83335, 1047343] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1047339, 237790] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1047339, 237790] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1040883, 86000] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1040883, 86000] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1026091, 544943] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1026091, 544943] processed_samples 1401 unjoint_samples 1400 joint_samples 3 [1047971, 167875] processed_samples 1401 unjoint_samples 1400 joint_samples 3 [1047971, 167875] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1039349, 289314] processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1039349, 289314] [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6cd7f7a80] stream 0, offset 0x101588b: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea94d7900] stream 0, offset 0x101588b: partial file [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1047339, 625503] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1047339, 625503] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1039349, 763921] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1039349, 763921] [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure processed_samples 1500 unjoint_samples 1500 joint_samples 3 [323527, 1037597] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [323527, 1037597] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [545061, 1047343] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [545061, 1047343] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1047538, 465769] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1047538, 465769] processed_samples 1501 unjoint_samples 1500 joint_samples 3 [1040883, 433662] processed_samples 1501 unjoint_samples 1500 joint_samples 3 [1047971, 477597] processed_samples 1501 unjoint_samples 1500 joint_samples 3 [1040883, 433662] processed_samples 1501 unjoint_samples 1500 joint_samples 3 [1047971, 477597] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1026091, 923291] processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1026091, 923291] [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea0f7dbc0] mmco: unref short failure [h264 @ 0x561ea0f7dbc0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1045296, 187556] processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1045296, 187556] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [870714, 1047343] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [870714, 1047343] [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1041464, 90546] processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1041464, 90546] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1047339, 1045489] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1047339, 1045489] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [677976, 1037597] processed_samples 1600 unjoint_samples 1600 joint_samples 3 [677976, 1037597] [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1047538, 814061] [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1047538, 814061] processed_samples 1601 unjoint_samples 1600 joint_samples 3 [1047971, 827491] processed_samples 1601 unjoint_samples 1600 joint_samples 3 [1047971, 827491] processed_samples 1601 unjoint_samples 1600 joint_samples 3 [1040883, 758515] processed_samples 1601 unjoint_samples 1600 joint_samples 3 [1040883, 758515] [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure processed_samples 1700 unjoint_samples 1700 joint_samples 4 [979157, 256280] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [979157, 256280] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [122966, 1046434] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [122966, 1046434] processed_samples 1701 unjoint_samples 1700 joint_samples 4 [69105, 1045145] processed_samples 1701 unjoint_samples 1700 joint_samples 4 [69105, 1045145] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1047339, 218324] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1047339, 218324] [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1041464, 445196] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1041464, 445196] [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1045296, 535052] processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1045296, 535052] processed_samples 1701 unjoint_samples 1700 joint_samples 4 [1047971, 128230] processed_samples 1701 unjoint_samples 1700 joint_samples 4 [1047971, 128230] processed_samples 1700 unjoint_samples 1700 joint_samples 3 [976318, 1037597] processed_samples 1700 unjoint_samples 1700 joint_samples 3 [976318, 1037597] [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1045113, 263492] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1045113, 263492] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [389479, 1046434] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [389479, 1046434] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [979157, 654664] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [979157, 654664] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1047339, 505364] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1047339, 505364] processed_samples 1801 unjoint_samples 1800 joint_samples 4 [1047971, 415280] processed_samples 1801 unjoint_samples 1800 joint_samples 4 [1047971, 415280] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1045296, 793652] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1045296, 793652] processed_samples 1801 unjoint_samples 1800 joint_samples 4 [363617, 1045145] processed_samples 1801 unjoint_samples 1800 joint_samples 4 [363617, 1045145] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1041464, 881462] processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1041464, 881462] [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1045113, 529162] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [775740, 1046434] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1047339, 779869] processed_samples 1900 unjoint_samples 1900 joint_samples 5 [1043602, 148115] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1047339, 779869] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1045113, 529162] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [775740, 1046434] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [979157, 916552] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [979157, 916552] processed_samples 1900 unjoint_samples 1900 joint_samples 5 [1043602, 148115] processed_samples 1901 unjoint_samples 1900 joint_samples 4 [691590, 1045145] processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1045296, 1024060] [h264 @ 0x561ea4492980] mmco: unref short failure processed_samples 1901 unjoint_samples 1900 joint_samples 4 [691590, 1045145] [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1045296, 1024060] [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 1901 unjoint_samples 1900 joint_samples 4 [1047971, 736008] processed_samples 1901 unjoint_samples 1900 joint_samples 4 [1047971, 736008] [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 2000 unjoint_samples 2000 joint_samples 5 [36467, 1032445] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [36467, 1032445] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [304033, 1045409] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [304033, 1045409] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1044846, 97389] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1044846, 97389] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1047987, 56626] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1047987, 56626] [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure processed_samples 2000 unjoint_samples 2000 joint_samples 4 [1045113, 793428] processed_samples 2000 unjoint_samples 2000 joint_samples 4 [1045113, 793428] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1043602, 468732] processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1043602, 468732] processed_samples 2001 unjoint_samples 2000 joint_samples 4 [1047971, 1006155] processed_samples 2001 unjoint_samples 2000 joint_samples 4 [1047971, 1006155] [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure processed_samples 2001 unjoint_samples 2000 joint_samples 4 [1039393, 1045145] processed_samples 2001 unjoint_samples 2000 joint_samples 4 [1039393, 1045145] [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure processed_samples 2100 unjoint_samples 2100 joint_samples 5 [281372, 1032445] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [281372, 1032445] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [62975, 1039730] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [62975, 1039730] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1043602, 710431] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1043602, 710431] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1047987, 364437] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1047987, 364437] processed_samples 2101 unjoint_samples 2100 joint_samples 5 [249364, 1047420] processed_samples 2101 unjoint_samples 2100 joint_samples 5 [249364, 1047420] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [668046, 1045409] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [668046, 1045409] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1044846, 397493] processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1044846, 397493] processed_samples 2101 unjoint_samples 2100 joint_samples 5 [1047971, 219257] processed_samples 2101 unjoint_samples 2100 joint_samples 5 [1047971, 219257] [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure processed_samples 2200 unjoint_samples 2200 joint_samples 5 [572007, 1032445] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [572007, 1032445] processed_samples 2200 unjoint_samples 2200 joint_samples 6 [1043602, 24679] processed_samples 2200 unjoint_samples 2200 joint_samples 6 [1043602, 24679] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [389349, 1039730] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [389349, 1039730] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1047987, 661852] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1047987, 661852] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1044846, 672642] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1044846, 672642] [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure processed_samples 2201 unjoint_samples 2200 joint_samples 5 [611401, 1047420] processed_samples 2201 unjoint_samples 2200 joint_samples 5 [611401, 1047420] processed_samples 2201 unjoint_samples 2200 joint_samples 5 [1047971, 491839] processed_samples 2201 unjoint_samples 2200 joint_samples 5 [1047971, 491839] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [916914, 1045409] processed_samples 2200 unjoint_samples 2200 joint_samples 5 [916914, 1045409] [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure processed_samples 2300 unjoint_samples 2300 joint_samples 6 [9937, 1046859] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [9937, 1046859] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1044080, 333409] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1044080, 333409] processed_samples 2301 unjoint_samples 2300 joint_samples 6 [82151, 1047420] processed_samples 2301 unjoint_samples 2300 joint_samples 6 [82151, 1047420] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1043602, 291515] processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1043602, 291515] processed_samples 2301 unjoint_samples 2300 joint_samples 5 [1047971, 768985] processed_samples 2300 unjoint_samples 2300 joint_samples 5 [919006, 1032445] processed_samples 2300 unjoint_samples 2300 joint_samples 5 [919006, 1032445] processed_samples 2301 unjoint_samples 2300 joint_samples 5 [1047971, 768985] processed_samples 2300 unjoint_samples 2300 joint_samples 5 [1044846, 954395] processed_samples 2300 unjoint_samples 2300 joint_samples 5 [1044846, 954395] processed_samples 2300 unjoint_samples 2300 joint_samples 5 [680948, 1039730] processed_samples 2300 unjoint_samples 2300 joint_samples 5 [680948, 1039730] [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure processed_samples 2400 unjoint_samples 2400 joint_samples 6 [316783, 1047157] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [316783, 1047157] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [139382, 1044019] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [139382, 1044019] [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure processed_samples 2400 unjoint_samples 2400 joint_samples 6 [311615, 1046859] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [311615, 1046859] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1044080, 905726] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1044080, 905726] processed_samples 2400 unjoint_samples 2400 joint_samples 5 [949958, 1039730] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1043602, 638207] processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1043602, 638207] processed_samples 2400 unjoint_samples 2400 joint_samples 5 [949958, 1039730] processed_samples 2401 unjoint_samples 2400 joint_samples 6 [386748, 1047420] processed_samples 2401 unjoint_samples 2400 joint_samples 6 [386748, 1047420] processed_samples 2401 unjoint_samples 2400 joint_samples 6 [157854, 1047252] processed_samples 2401 unjoint_samples 2400 joint_samples 6 [157854, 1047252] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6ce62ae40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure processed_samples 2501 unjoint_samples 2500 joint_samples 6 [383377, 1047252] processed_samples 2501 unjoint_samples 2500 joint_samples 6 [383377, 1047252] processed_samples 2500 unjoint_samples 2500 joint_samples 7 [108867, 1044676] processed_samples 2500 unjoint_samples 2500 joint_samples 7 [108867, 1044676] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1048303, 181826] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1048303, 181826] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [436053, 1044019] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [701203, 1046859] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [436053, 1044019] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [701203, 1046859] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [624192, 1047157] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [624192, 1047157] processed_samples 2501 unjoint_samples 2500 joint_samples 6 [680517, 1047420] processed_samples 2501 unjoint_samples 2500 joint_samples 6 [680517, 1047420] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1043602, 1030183] processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1043602, 1030183] [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure /root/miniconda3/envs/py38/lib/python3.8/site-packages/PIL/TiffImagePlugin.py:870: UserWarning: Corrupt EXIF data. Expecting to read 4 bytes but only got 0. warnings.warn(str(msg)) /root/miniconda3/envs/py38/lib/python3.8/site-packages/PIL/TiffImagePlugin.py:870: UserWarning: Corrupt EXIF data. Expecting to read 4 bytes but only got 0. warnings.warn(str(msg)) [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d5e5dc40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d5e5dc40] mmco: unref short failure [h264 @ 0x55e6d5e5dc40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure processed_samples 2600 unjoint_samples 2600 joint_samples 7 [417591, 1044676] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [417591, 1044676] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [821080, 1044019] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [54690, 1046859] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [54690, 1046859] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [821080, 1044019] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [278341, 1045441] processed_samples 2600 unjoint_samples 2600 joint_samples 7 [278341, 1045441] processed_samples 2601 unjoint_samples 2600 joint_samples 6 [687210, 1047252] processed_samples 2601 unjoint_samples 2600 joint_samples 6 [687210, 1047252] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1048303, 613581] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1048303, 613581] processed_samples 2601 unjoint_samples 2600 joint_samples 7 [32427, 1047478] processed_samples 2601 unjoint_samples 2600 joint_samples 7 [32427, 1047478] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [947706, 1047157] processed_samples 2600 unjoint_samples 2600 joint_samples 6 [947706, 1047157] [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure processed_samples 2700 unjoint_samples 2700 joint_samples 7 [137915, 1046746] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [137915, 1046746] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [758058, 1044676] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [758058, 1044676] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [341757, 1046859] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [341757, 1046859] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [563205, 1045441] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [563205, 1045441] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1027164, 246787] processed_samples 2700 unjoint_samples 2700 joint_samples 6 [1048303, 915905] processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1027164, 246787] processed_samples 2700 unjoint_samples 2700 joint_samples 6 [1048303, 915905] processed_samples 2701 unjoint_samples 2700 joint_samples 7 [379538, 1047478] [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 2701 unjoint_samples 2700 joint_samples 7 [379538, 1047478] [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure processed_samples 2701 unjoint_samples 2700 joint_samples 7 [2253, 1047252] processed_samples 2701 unjoint_samples 2700 joint_samples 7 [2253, 1047252] [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1027164, 568592] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [123626, 1046952] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1027164, 568592] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [123626, 1046952] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [379112, 1046746] processed_samples 2801 unjoint_samples 2800 joint_samples 7 [243258, 1047252] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [379112, 1046746] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [892132, 1046859] processed_samples 2801 unjoint_samples 2800 joint_samples 7 [243258, 1047252] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [892132, 1046859] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1034834, 1044676] [h264 @ 0x561ea15a4780] mmco: unref short failure processed_samples 2800 unjoint_samples 2800 joint_samples 7 [809730, 1045441] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1034834, 1044676] [h264 @ 0x55e6d1810ac0] mmco: unref short failure processed_samples 2801 unjoint_samples 2800 joint_samples 7 [715231, 1047478] processed_samples 2800 unjoint_samples 2800 joint_samples 7 [809730, 1045441] processed_samples 2801 unjoint_samples 2800 joint_samples 7 [715231, 1047478] [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1044485, 290540] processed_samples 2901 unjoint_samples 2900 joint_samples 7 [624439, 1047252] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1044485, 290540] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1046735, 203578] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1046735, 203578] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [383200, 1046952] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [383200, 1046952] processed_samples 2901 unjoint_samples 2900 joint_samples 7 [624439, 1047252] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [621932, 1046746] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [621932, 1046746] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1027164, 874454] processed_samples 2901 unjoint_samples 2900 joint_samples 7 [953565, 1047478] processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1027164, 874454] processed_samples 2901 unjoint_samples 2900 joint_samples 7 [953565, 1047478] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1047093, 2273] processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1047093, 2273] [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1046735, 561902] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1046735, 561902] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [18394, 1046674] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [18394, 1046674] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1044485, 578516] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1044485, 578516] [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1047093, 457707] processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1047093, 457707] processed_samples 3000 unjoint_samples 3000 joint_samples 7 [959805, 1046746] processed_samples 3000 unjoint_samples 3000 joint_samples 7 [959805, 1046746] processed_samples 3001 unjoint_samples 3000 joint_samples 8 [177339, 1047478] processed_samples 3001 unjoint_samples 3000 joint_samples 8 [177339, 1047478] processed_samples 3001 unjoint_samples 3000 joint_samples 7 [972063, 1047252] processed_samples 3001 unjoint_samples 3000 joint_samples 7 [972063, 1047252] [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure processed_samples 3000 unjoint_samples 3000 joint_samples 7 [703810, 1046952] processed_samples 3000 unjoint_samples 3000 joint_samples 7 [703810, 1046952] [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1019452, 301782] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1019452, 301782] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [276455, 1046674] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [276455, 1046674] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1046735, 809227] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1046735, 809227] processed_samples 3101 unjoint_samples 3100 joint_samples 8 [1046835, 261761] [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure processed_samples 3101 unjoint_samples 3100 joint_samples 8 [1046835, 261761] [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1047093, 649298] processed_samples 3101 unjoint_samples 3100 joint_samples 8 [466325, 1047478] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1047093, 649298] processed_samples 3101 unjoint_samples 3100 joint_samples 8 [466325, 1047478] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1044485, 925438] processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1044485, 925438] [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure processed_samples 3100 unjoint_samples 3100 joint_samples 7 [1010890, 1046952] processed_samples 3100 unjoint_samples 3100 joint_samples 7 [1010890, 1046952] [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure processed_samples 3200 unjoint_samples 3200 joint_samples 8 [240095, 1047144] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [240095, 1047144] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [1044485, 126914] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [1044485, 126914] processed_samples 3201 unjoint_samples 3200 joint_samples 8 [1046835, 628658] processed_samples 3201 unjoint_samples 3200 joint_samples 8 [1046835, 628658] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [1047346, 113318] processed_samples 3200 unjoint_samples 3200 joint_samples 9 [1047346, 113318] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1019452, 673252] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1019452, 673252] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [604093, 1046674] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [604093, 1046674] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1047093, 928239] processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1047093, 928239] [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure processed_samples 3201 unjoint_samples 3200 joint_samples 8 [819044, 1047478] processed_samples 3201 unjoint_samples 3200 joint_samples 8 [819044, 1047478] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cddb1e40] mmco: unref short failure [h264 @ 0x55e6cddb1e40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1044485, 431862] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1044485, 431862] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1047346, 376573] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1047346, 376573] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1047093, 136582] processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1047093, 136582] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [878470, 1046674] processed_samples 3301 unjoint_samples 3300 joint_samples 9 [1041086, 31893] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [878470, 1046674] processed_samples 3301 unjoint_samples 3300 joint_samples 9 [1041086, 31893] [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 3301 unjoint_samples 3300 joint_samples 8 [1046835, 992113] processed_samples 3301 unjoint_samples 3300 joint_samples 8 [1046835, 992113] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [501028, 1047144] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [501028, 1047144] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [1019452, 963737] processed_samples 3300 unjoint_samples 3300 joint_samples 8 [1019452, 963737] [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cddb1e40] mmco: unref short failure [h264 @ 0x55e6cddb1e40] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x55e6cddb1e40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1039036, 137212] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1047346, 741948] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1039036, 137212] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1047346, 741948] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [220598, 1041356] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [220598, 1041356] processed_samples 3401 unjoint_samples 3400 joint_samples 9 [1046835, 177431] processed_samples 3401 unjoint_samples 3400 joint_samples 9 [1046835, 177431] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1044485, 695164] [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1044485, 695164] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1047093, 554613] processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1047093, 554613] processed_samples 3401 unjoint_samples 3400 joint_samples 9 [1041086, 283922] processed_samples 3401 unjoint_samples 3400 joint_samples 9 [1041086, 283922] processed_samples 3400 unjoint_samples 3400 joint_samples 8 [780195, 1047144] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure processed_samples 3400 unjoint_samples 3400 joint_samples 8 [780195, 1047144] [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure processed_samples 3500 unjoint_samples 3500 joint_samples 9 [728645, 1041356] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [728645, 1041356] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1039036, 407682] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1039036, 407682] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1044485, 933891] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1044485, 933891] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1047346, 1012686] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1047346, 1012686] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1032750, 32133] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1032750, 32133] processed_samples 3501 unjoint_samples 3500 joint_samples 9 [1046835, 648663] processed_samples 3501 unjoint_samples 3500 joint_samples 9 [1041086, 620752] processed_samples 3501 unjoint_samples 3500 joint_samples 9 [1041086, 620752] processed_samples 3501 unjoint_samples 3500 joint_samples 9 [1046835, 648663] [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1047093, 853556] processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1047093, 853556] [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure processed_samples 3600 unjoint_samples 3600 joint_samples 10 [231537, 1045544] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [231537, 1045544] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [214630, 1045586] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [214630, 1045586] [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 3600 unjoint_samples 3600 joint_samples 9 [980620, 1041356] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [980620, 1041356] [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure processed_samples 3600 unjoint_samples 3600 joint_samples 10 [328909, 1034998] processed_samples 3600 unjoint_samples 3600 joint_samples 10 [328909, 1034998] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [1039036, 669418] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [1039036, 669418] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [1032750, 295506] processed_samples 3600 unjoint_samples 3600 joint_samples 9 [1032750, 295506] processed_samples 3601 unjoint_samples 3600 joint_samples 9 [1041086, 885198] processed_samples 3601 unjoint_samples 3600 joint_samples 9 [1041086, 885198] processed_samples 3601 unjoint_samples 3600 joint_samples 10 [40771, 1010669] processed_samples 3601 unjoint_samples 3600 joint_samples 10 [40771, 1010669] [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure processed_samples 3700 unjoint_samples 3700 joint_samples 10 [516091, 1045544] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [516091, 1045544] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1029442, 309517] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1029442, 309517] [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure processed_samples 3701 unjoint_samples 3700 joint_samples 10 [322262, 1010669] processed_samples 3701 unjoint_samples 3700 joint_samples 10 [322262, 1010669] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [526620, 1045586] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [526620, 1045586] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [705231, 1034998] processed_samples 3700 unjoint_samples 3700 joint_samples 10 [705231, 1034998] processed_samples 3701 unjoint_samples 3700 joint_samples 10 [1045430, 100224] processed_samples 3701 unjoint_samples 3700 joint_samples 10 [1045430, 100224] processed_samples 3700 unjoint_samples 3700 joint_samples 9 [1032750, 624494] processed_samples 3700 unjoint_samples 3700 joint_samples 9 [1032750, 624494] processed_samples 3700 unjoint_samples 3700 joint_samples 9 [1039036, 926072] processed_samples 3700 unjoint_samples 3700 joint_samples 9 [1039036, 926072] [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1041143, 141369] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1041143, 141369] [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1029442, 609909] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [958748, 1034998] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [958748, 1034998] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1029442, 609909] processed_samples 3801 unjoint_samples 3800 joint_samples 10 [1045430, 385647] processed_samples 3801 unjoint_samples 3800 joint_samples 10 [1045430, 385647] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [889337, 1045544] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [889337, 1045544] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [877724, 1045586] processed_samples 3800 unjoint_samples 3800 joint_samples 10 [877724, 1045586] processed_samples 3800 unjoint_samples 3800 joint_samples 9 [1032750, 962281] processed_samples 3800 unjoint_samples 3800 joint_samples 9 [1032750, 962281] processed_samples 3801 unjoint_samples 3800 joint_samples 10 [586915, 1010669] [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure processed_samples 3801 unjoint_samples 3800 joint_samples 10 [586915, 1010669] [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure processed_samples 3900 unjoint_samples 3900 joint_samples 10 [175163, 1044984] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [175163, 1044984] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1041331, 340120] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1045066, 166543] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1045066, 166543] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1041331, 340120] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [189276, 1045586] processed_samples 3900 unjoint_samples 3900 joint_samples 11 [189276, 1045586] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1041143, 578141] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1041143, 578141] processed_samples 3901 unjoint_samples 3900 joint_samples 10 [1045430, 752033] processed_samples 3901 unjoint_samples 3900 joint_samples 10 [1045430, 752033] processed_samples 3901 unjoint_samples 3900 joint_samples 10 [924615, 1010669] processed_samples 3901 unjoint_samples 3900 joint_samples 10 [924615, 1010669] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1029442, 870023] processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1029442, 870023] [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4bdeec0] [h264 @ 0x55e6d4750cc0] mmco: unref short failure mmco: unref short failure [h264 @ 0x561ea4bdeec0] [h264 @ 0x55e6d4750cc0] mmco: unref short failure mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1045066, 443885] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [237007, 1029479] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1045066, 443885] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [237007, 1029479] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1041331, 538356] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1041331, 538356] processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1041143, 924587] [h264 @ 0x55e6ce2c0c40] mmco: unref short failure processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1041143, 924587] processed_samples 4001 unjoint_samples 4000 joint_samples 11 [95693, 1030749] processed_samples 4001 unjoint_samples 4000 joint_samples 11 [95693, 1030749] [h264 @ 0x561ea5f27180] mmco: unref short failure processed_samples 4000 unjoint_samples 4000 joint_samples 11 [480306, 1045586] processed_samples 4000 unjoint_samples 4000 joint_samples 10 [459361, 1044984] processed_samples 4000 unjoint_samples 4000 joint_samples 11 [480306, 1045586] processed_samples 4000 unjoint_samples 4000 joint_samples 10 [459361, 1044984] processed_samples 4001 unjoint_samples 4000 joint_samples 11 [1047632, 31448] processed_samples 4001 unjoint_samples 4000 joint_samples 11 [1047632, 31448] [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea5f27180] mmco: unref short failure [h264 @ 0x561ea5f27180] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure processed_samples 4100 unjoint_samples 4100 joint_samples 11 [483810, 1029479] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [483810, 1029479] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [350782, 1038921] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [350782, 1038921] processed_samples 4100 unjoint_samples 4100 joint_samples 10 [733097, 1044984] processed_samples 4100 unjoint_samples 4100 joint_samples 10 [733097, 1044984] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [809266, 1045586] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [809266, 1045586] processed_samples 4101 unjoint_samples 4100 joint_samples 11 [375393, 1030749] processed_samples 4101 unjoint_samples 4100 joint_samples 11 [375393, 1030749] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1041331, 833898] processed_samples 4101 unjoint_samples 4100 joint_samples 11 [1047632, 300672] processed_samples 4101 unjoint_samples 4100 joint_samples 11 [1047632, 300672] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1041331, 833898] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1045066, 720053] processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1045066, 720053] [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea2aadf40] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure processed_samples 4200 unjoint_samples 4200 joint_samples 12 [65712, 1041530] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [65712, 1041530] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [1047824, 115756] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [1047824, 115756] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [762572, 1029479] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [643982, 1038921] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [762572, 1029479] processed_samples 4200 unjoint_samples 4200 joint_samples 11 [643982, 1038921] processed_samples 4201 unjoint_samples 4200 joint_samples 11 [741625, 1030749] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [1047445, 278634] processed_samples 4201 unjoint_samples 4200 joint_samples 11 [741625, 1030749] processed_samples 4200 unjoint_samples 4200 joint_samples 12 [1047445, 278634] processed_samples 4201 unjoint_samples 4200 joint_samples 11 [1047632, 657953] processed_samples 4201 unjoint_samples 4200 joint_samples 11 [1047632, 657953] processed_samples 4200 unjoint_samples 4200 joint_samples 10 [1004773, 1044984] processed_samples 4200 unjoint_samples 4200 joint_samples 10 [1004773, 1044984] [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure processed_samples 4300 unjoint_samples 4300 joint_samples 11 [1046875, 219662] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [357576, 1041530] processed_samples 4300 unjoint_samples 4300 joint_samples 11 [1046875, 219662] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [357576, 1041530] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1047824, 391700] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1047824, 391700] [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure processed_samples 4300 unjoint_samples 4300 joint_samples 11 [947015, 1038921] processed_samples 4300 unjoint_samples 4300 joint_samples 11 [947015, 1038921] [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1047445, 506995] processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1047445, 506995] processed_samples 4300 unjoint_samples 4300 joint_samples 11 [1044773, 1046301] processed_samples 4300 unjoint_samples 4300 joint_samples 11 [1044773, 1046301] processed_samples 4301 unjoint_samples 4300 joint_samples 11 [1029906, 1030749] processed_samples 4301 unjoint_samples 4300 joint_samples 11 [1029906, 1030749] processed_samples 4301 unjoint_samples 4300 joint_samples 12 [8015, 1047437] processed_samples 4301 unjoint_samples 4300 joint_samples 12 [8015, 1047437] [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure processed_samples 4400 unjoint_samples 4400 joint_samples 12 [317522, 1048202] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [317522, 1048202] [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1046476, 202506] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1046476, 202506] processed_samples 4400 unjoint_samples 4400 joint_samples 11 [1046875, 575594] processed_samples 4400 unjoint_samples 4400 joint_samples 11 [1046875, 575594] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1047824, 724016] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [692091, 1041530] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1047824, 724016] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [692091, 1041530] processed_samples 4401 unjoint_samples 4400 joint_samples 12 [230763, 1046101] processed_samples 4401 unjoint_samples 4400 joint_samples 12 [321746, 1047437] processed_samples 4401 unjoint_samples 4400 joint_samples 12 [230763, 1046101] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1047445, 812157] processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1047445, 812157] processed_samples 4401 unjoint_samples 4400 joint_samples 12 [321746, 1047437] [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure processed_samples 4500 unjoint_samples 4500 joint_samples 13 [85163, 1046856] processed_samples 4500 unjoint_samples 4500 joint_samples 13 [85163, 1046856] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [590164, 1048202] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [590164, 1048202] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1047824, 1005832] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1047824, 1005832] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1046476, 551218] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1046476, 551218] processed_samples 4500 unjoint_samples 4500 joint_samples 11 [1046875, 890870] processed_samples 4500 unjoint_samples 4500 joint_samples 11 [1046875, 890870] processed_samples 4501 unjoint_samples 4500 joint_samples 12 [518848, 1046101] processed_samples 4501 unjoint_samples 4500 joint_samples 12 [518848, 1046101] processed_samples 4501 unjoint_samples 4500 joint_samples 12 [663225, 1047437] processed_samples 4501 unjoint_samples 4500 joint_samples 12 [663225, 1047437] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1045795, 1045356] processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1045795, 1045356] [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure processed_samples 4600 unjoint_samples 4600 joint_samples 13 [256407, 1045356] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [256407, 1045356] [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1047824, 254905] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1047824, 254905] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [376180, 1046856] processed_samples 4600 unjoint_samples 4600 joint_samples 13 [376180, 1046856] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [162696, 1043538] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [162696, 1043538] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [859240, 1048202] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [859240, 1048202] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [1046476, 786876] processed_samples 4600 unjoint_samples 4600 joint_samples 12 [1046476, 786876] processed_samples 4601 unjoint_samples 4600 joint_samples 12 [815658, 1046101] processed_samples 4601 unjoint_samples 4600 joint_samples 12 [815658, 1046101] processed_samples 4601 unjoint_samples 4600 joint_samples 12 [988397, 1047437] processed_samples 4601 unjoint_samples 4600 joint_samples 12 [988397, 1047437] [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561e7e109280] stream 1, offset 0x14000d8: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6d163f280] stream 1, offset 0x14000d8: partial file [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1046476, 26108] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1046476, 26108] processed_samples 4701 unjoint_samples 4700 joint_samples 13 [1014380, 131253] processed_samples 4701 unjoint_samples 4700 joint_samples 13 [1014380, 131253] processed_samples 4701 unjoint_samples 4700 joint_samples 12 [567554, 1043538] processed_samples 4701 unjoint_samples 4700 joint_samples 12 [567554, 1043538] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [612562, 1045356] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [612562, 1045356] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047593, 36467] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047593, 36467] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [606091, 1046856] processed_samples 4701 unjoint_samples 4700 joint_samples 13 [1046186, 255012] processed_samples 4701 unjoint_samples 4700 joint_samples 13 [1046186, 255012] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [606091, 1046856] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047824, 677000] processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047824, 677000] [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1047593, 319966] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1047593, 319966] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1047824, 972785] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1047824, 972785] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1046476, 408336] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1046476, 408336] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [954253, 1046856] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [954253, 1046856] processed_samples 4801 unjoint_samples 4800 joint_samples 13 [1046186, 512793] processed_samples 4801 unjoint_samples 4800 joint_samples 12 [874702, 1043538] processed_samples 4801 unjoint_samples 4800 joint_samples 13 [1046186, 512793] processed_samples 4801 unjoint_samples 4800 joint_samples 12 [874702, 1043538] processed_samples 4801 unjoint_samples 4800 joint_samples 13 [1014380, 459778] processed_samples 4801 unjoint_samples 4800 joint_samples 13 [1014380, 459778] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1027010, 1045356] processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1027010, 1045356] [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure processed_samples 4900 unjoint_samples 4900 joint_samples 14 [235792, 1045412] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [235792, 1045412] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [267869, 1031128] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [267869, 1031128] processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1047593, 691149] processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1047593, 691149] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [1024519, 307876] processed_samples 4900 unjoint_samples 4900 joint_samples 14 [1024519, 307876] processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1046476, 716206] processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1046476, 716206] processed_samples 4901 unjoint_samples 4900 joint_samples 13 [1047690, 107218] processed_samples 4901 unjoint_samples 4900 joint_samples 13 [1047690, 107218] processed_samples 4901 unjoint_samples 4900 joint_samples 13 [1014380, 785409] processed_samples 4901 unjoint_samples 4900 joint_samples 13 [1014380, 785409] processed_samples 4901 unjoint_samples 4900 joint_samples 13 [1046186, 902312] processed_samples 4901 unjoint_samples 4900 joint_samples 13 [1046186, 902312] [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1046476, 90511] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1046476, 90511] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1047593, 46292] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1047593, 46292] processed_samples 5001 unjoint_samples 5000 joint_samples 14 [65112, 1011428] [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure processed_samples 5001 unjoint_samples 5000 joint_samples 14 [65112, 1011428] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [575559, 1045412] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [575559, 1045412] processed_samples 5001 unjoint_samples 5000 joint_samples 14 [164226, 1042734] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [604960, 1031128] processed_samples 5001 unjoint_samples 5000 joint_samples 14 [164226, 1042734] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [604960, 1031128] [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1024519, 646975] processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1024519, 646975] processed_samples 5001 unjoint_samples 5000 joint_samples 13 [1047690, 381393] processed_samples 5001 unjoint_samples 5000 joint_samples 13 [1047690, 381393] [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea27fab00] mmco: unref short failure [h264 @ 0x561ea27fab00] mmco: unref short failure [h264 @ 0x561ea27fab00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1047593, 366348] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1047593, 366348] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1046476, 375029] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1046476, 375029] processed_samples 5101 unjoint_samples 5100 joint_samples 14 [403103, 1011428] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [906178, 1045412] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [906178, 1045412] processed_samples 5101 unjoint_samples 5100 joint_samples 14 [403103, 1011428] [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1030343, 1030990] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [1030343, 1030990] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [930489, 1031128] processed_samples 5101 unjoint_samples 5100 joint_samples 13 [1047690, 619643] processed_samples 5101 unjoint_samples 5100 joint_samples 14 [521507, 1042734] processed_samples 5101 unjoint_samples 5100 joint_samples 13 [1047690, 619643] processed_samples 5100 unjoint_samples 5100 joint_samples 14 [930489, 1031128] processed_samples 5101 unjoint_samples 5100 joint_samples 14 [521507, 1042734] [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1039191, 218605] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1039191, 218605] [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure processed_samples 5200 unjoint_samples 5200 joint_samples 14 [1046476, 633736] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [293629, 1031128] [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure processed_samples 5200 unjoint_samples 5200 joint_samples 14 [1046476, 633736] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [293629, 1031128] processed_samples 5200 unjoint_samples 5200 joint_samples 14 [1047593, 575923] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [409402, 1046384] processed_samples 5200 unjoint_samples 5200 joint_samples 14 [1047593, 575923] processed_samples 5200 unjoint_samples 5200 joint_samples 15 [409402, 1046384] processed_samples 5201 unjoint_samples 5200 joint_samples 14 [714111, 1011428] [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure processed_samples 5201 unjoint_samples 5200 joint_samples 14 [714111, 1011428] [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure processed_samples 5201 unjoint_samples 5200 joint_samples 14 [851489, 1042734] processed_samples 5201 unjoint_samples 5200 joint_samples 14 [851489, 1042734] processed_samples 5201 unjoint_samples 5200 joint_samples 13 [1047690, 939144] processed_samples 5201 unjoint_samples 5200 joint_samples 13 [1047690, 939144] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdab6f80] mmco: unref short failure [h264 @ 0x55e6cdab6f80] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure processed_samples 5300 unjoint_samples 5300 joint_samples 14 [1047593, 943718] processed_samples 5300 unjoint_samples 5300 joint_samples 14 [1047593, 943718] [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1039191, 523024] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1039191, 523024] processed_samples 5301 unjoint_samples 5300 joint_samples 14 [206323, 1047298] [h264 @ 0x55e6ce443040] mmco: unref short failure processed_samples 5301 unjoint_samples 5300 joint_samples 14 [206323, 1047298] [h264 @ 0x561ea4ce19c0] mmco: unref short failure processed_samples 5300 unjoint_samples 5300 joint_samples 15 [863799, 1046384] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [863799, 1046384] processed_samples 5301 unjoint_samples 5300 joint_samples 15 [1014188, 223190] processed_samples 5301 unjoint_samples 5300 joint_samples 15 [1014188, 223190] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [601513, 1031128] processed_samples 5300 unjoint_samples 5300 joint_samples 15 [601513, 1031128] processed_samples 5300 unjoint_samples 5300 joint_samples 14 [1046476, 903227] processed_samples 5300 unjoint_samples 5300 joint_samples 14 [1046476, 903227] processed_samples 5301 unjoint_samples 5300 joint_samples 14 [962646, 1011428] [h264 @ 0x55e6ce9550c0] mmco: unref short failure processed_samples 5301 unjoint_samples 5300 joint_samples 14 [962646, 1011428] [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure processed_samples 5400 unjoint_samples 5400 joint_samples 15 [108055, 1048003] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [108055, 1048003] processed_samples 5401 unjoint_samples 5400 joint_samples 15 [1029376, 200376] processed_samples 5401 unjoint_samples 5400 joint_samples 15 [1029376, 200376] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [261656, 1039727] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [261656, 1039727] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1039191, 851814] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1039191, 851814] processed_samples 5400 unjoint_samples 5400 joint_samples 16 [175795, 1046475] processed_samples 5400 unjoint_samples 5400 joint_samples 16 [175795, 1046475] processed_samples 5401 unjoint_samples 5400 joint_samples 15 [1014188, 530602] processed_samples 5401 unjoint_samples 5400 joint_samples 15 [1014188, 530602] processed_samples 5401 unjoint_samples 5400 joint_samples 14 [490678, 1047298] processed_samples 5401 unjoint_samples 5400 joint_samples 14 [490678, 1047298] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [908335, 1031128] processed_samples 5400 unjoint_samples 5400 joint_samples 15 [908335, 1031128] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea07202c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure processed_samples 5500 unjoint_samples 5500 joint_samples 16 [1045276, 135972] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [1045276, 135972] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [142239, 1047149] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [142239, 1047149] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [349672, 1048003] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [349672, 1048003] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [565488, 1039727] processed_samples 5500 unjoint_samples 5500 joint_samples 15 [565488, 1039727] processed_samples 5501 unjoint_samples 5500 joint_samples 14 [841654, 1047298] processed_samples 5501 unjoint_samples 5500 joint_samples 14 [841654, 1047298] processed_samples 5501 unjoint_samples 5500 joint_samples 15 [1029376, 698000] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [565048, 1046475] processed_samples 5501 unjoint_samples 5500 joint_samples 15 [1029376, 698000] processed_samples 5500 unjoint_samples 5500 joint_samples 16 [565048, 1046475] processed_samples 5501 unjoint_samples 5500 joint_samples 15 [1014188, 875717] processed_samples 5501 unjoint_samples 5500 joint_samples 15 [1014188, 875717] [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure processed_samples 5600 unjoint_samples 5600 joint_samples 16 [911603, 1046475] [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure processed_samples 5600 unjoint_samples 5600 joint_samples 16 [911603, 1046475] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure processed_samples 5600 unjoint_samples 5600 joint_samples 15 [607188, 1048003] processed_samples 5600 unjoint_samples 5600 joint_samples 15 [607188, 1048003] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1045276, 353542] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1045276, 353542] processed_samples 5601 unjoint_samples 5600 joint_samples 16 [159906, 1042949] processed_samples 5601 unjoint_samples 5600 joint_samples 15 [187110, 1047298] processed_samples 5601 unjoint_samples 5600 joint_samples 16 [159906, 1042949] processed_samples 5601 unjoint_samples 5600 joint_samples 15 [187110, 1047298] [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure processed_samples 5600 unjoint_samples 5600 joint_samples 16 [465498, 1047149] processed_samples 5600 unjoint_samples 5600 joint_samples 16 [465498, 1047149] processed_samples 5601 unjoint_samples 5600 joint_samples 15 [1039756, 1040344] processed_samples 5601 unjoint_samples 5600 joint_samples 15 [1039756, 1040344] processed_samples 5600 unjoint_samples 5600 joint_samples 15 [850381, 1039727] processed_samples 5600 unjoint_samples 5600 joint_samples 15 [850381, 1039727] [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1045837, 85506] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1045837, 85506] processed_samples 5700 unjoint_samples 5700 joint_samples 17 [1043760, 209231] processed_samples 5700 unjoint_samples 5700 joint_samples 17 [1043760, 209231] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1045276, 652416] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1045276, 652416] processed_samples 5701 unjoint_samples 5700 joint_samples 16 [373400, 1047410] processed_samples 5701 unjoint_samples 5700 joint_samples 16 [373400, 1047410] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [729141, 1047149] processed_samples 5701 unjoint_samples 5700 joint_samples 16 [597950, 1042949] processed_samples 5700 unjoint_samples 5700 joint_samples 16 [729141, 1047149] processed_samples 5701 unjoint_samples 5700 joint_samples 16 [597950, 1042949] processed_samples 5701 unjoint_samples 5700 joint_samples 15 [484423, 1047298] processed_samples 5701 unjoint_samples 5700 joint_samples 15 [484423, 1047298] processed_samples 5700 unjoint_samples 5700 joint_samples 15 [827293, 1048003] processed_samples 5700 unjoint_samples 5700 joint_samples 15 [827293, 1048003] [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea5f27180] mmco: unref short failure processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1045837, 362580] processed_samples 5800 unjoint_samples 5800 joint_samples 17 [1046817, 52480] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1045837, 362580] processed_samples 5800 unjoint_samples 5800 joint_samples 17 [1046817, 52480] [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 5800 unjoint_samples 5800 joint_samples 17 [1043760, 490702] processed_samples 5800 unjoint_samples 5800 joint_samples 17 [1043760, 490702] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1045276, 974048] processed_samples 5801 unjoint_samples 5800 joint_samples 16 [667760, 1047410] processed_samples 5801 unjoint_samples 5800 joint_samples 16 [667760, 1047410] processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1045276, 974048] processed_samples 5801 unjoint_samples 5800 joint_samples 16 [871812, 1042949] processed_samples 5801 unjoint_samples 5800 joint_samples 16 [871812, 1042949] processed_samples 5801 unjoint_samples 5800 joint_samples 15 [807929, 1047298] processed_samples 5801 unjoint_samples 5800 joint_samples 15 [807929, 1047298] processed_samples 5800 unjoint_samples 5800 joint_samples 15 [1047623, 1048003] processed_samples 5800 unjoint_samples 5800 joint_samples 15 [1047623, 1048003] [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cec12f80] mmco: unref short failure [h264 @ 0x55e6cec12f80] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1045837, 623962] processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1045837, 623962] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [1046817, 312370] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [1046817, 312370] [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1047623, 261924] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [311983, 1047110] processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1047623, 261924] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [311983, 1047110] processed_samples 5901 unjoint_samples 5900 joint_samples 16 [1023756, 47536] processed_samples 5901 unjoint_samples 5900 joint_samples 16 [1023756, 47536] processed_samples 5901 unjoint_samples 5900 joint_samples 16 [923554, 1047410] processed_samples 5901 unjoint_samples 5900 joint_samples 16 [923554, 1047410] processed_samples 5901 unjoint_samples 5900 joint_samples 17 [1035738, 110992] processed_samples 5901 unjoint_samples 5900 joint_samples 17 [1035738, 110992] [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure processed_samples 5900 unjoint_samples 5900 joint_samples 17 [1043760, 792785] processed_samples 5900 unjoint_samples 5900 joint_samples 17 [1043760, 792785] [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1046817, 618235] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1046817, 618235] processed_samples 6000 unjoint_samples 6000 joint_samples 18 [46850, 1037237] processed_samples 6000 unjoint_samples 6000 joint_samples 18 [46850, 1037237] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [667161, 1047110] processed_samples 6000 unjoint_samples 6000 joint_samples 17 [667161, 1047110] processed_samples 6000 unjoint_samples 6000 joint_samples 16 [1047623, 508245] processed_samples 6000 unjoint_samples 6000 joint_samples 16 [1047623, 508245] processed_samples 6001 unjoint_samples 6000 joint_samples 17 [1046010, 195616] processed_samples 6001 unjoint_samples 6000 joint_samples 17 [1046010, 195616] processed_samples 6001 unjoint_samples 6000 joint_samples 17 [1035738, 433490] processed_samples 6001 unjoint_samples 6000 joint_samples 16 [1023756, 297290] processed_samples 6001 unjoint_samples 6000 joint_samples 17 [1035738, 433490] processed_samples 6001 unjoint_samples 6000 joint_samples 16 [1023756, 297290] processed_samples 6000 unjoint_samples 6000 joint_samples 16 [1045837, 869868] processed_samples 6000 unjoint_samples 6000 joint_samples 16 [1045837, 869868] [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 6100 unjoint_samples 6100 joint_samples 17 [91569, 1045044] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [91569, 1045044] processed_samples 6100 unjoint_samples 6100 joint_samples 18 [459519, 1037237] processed_samples 6100 unjoint_samples 6100 joint_samples 18 [459519, 1037237] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [976889, 1047110] processed_samples 6100 unjoint_samples 6100 joint_samples 16 [1047623, 765236] processed_samples 6101 unjoint_samples 6100 joint_samples 17 [1046010, 440785] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [976889, 1047110] processed_samples 6101 unjoint_samples 6100 joint_samples 16 [1023756, 584187] processed_samples 6100 unjoint_samples 6100 joint_samples 16 [1047623, 765236] processed_samples 6101 unjoint_samples 6100 joint_samples 17 [1046010, 440785] processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1046817, 938909] processed_samples 6101 unjoint_samples 6100 joint_samples 16 [1023756, 584187] [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1046817, 938909] processed_samples 6101 unjoint_samples 6100 joint_samples 17 [1035738, 980877] [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure processed_samples 6101 unjoint_samples 6100 joint_samples 17 [1035738, 980877] [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure processed_samples 6200 unjoint_samples 6200 joint_samples 17 [104603, 1041805] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [104603, 1041805] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [158147, 1033978] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [158147, 1033978] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [359209, 1045044] processed_samples 6200 unjoint_samples 6200 joint_samples 17 [359209, 1045044] [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure processed_samples 6200 unjoint_samples 6200 joint_samples 18 [236439, 1047110] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [236439, 1047110] processed_samples 6201 unjoint_samples 6200 joint_samples 18 [359253, 1005743] processed_samples 6201 unjoint_samples 6200 joint_samples 18 [359253, 1005743] processed_samples 6201 unjoint_samples 6200 joint_samples 17 [1046010, 755565] processed_samples 6201 unjoint_samples 6200 joint_samples 17 [1046010, 755565] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [797105, 1037237] processed_samples 6200 unjoint_samples 6200 joint_samples 18 [797105, 1037237] processed_samples 6201 unjoint_samples 6200 joint_samples 16 [1023756, 958362] processed_samples 6201 unjoint_samples 6200 joint_samples 16 [1023756, 958362] [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure processed_samples 6300 unjoint_samples 6300 joint_samples 18 [480000, 1033978] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [480000, 1033978] processed_samples 6300 unjoint_samples 6300 joint_samples 17 [635506, 1045044] processed_samples 6300 unjoint_samples 6300 joint_samples 17 [635506, 1045044] processed_samples 6300 unjoint_samples 6300 joint_samples 19 [1046712, 70047] processed_samples 6300 unjoint_samples 6300 joint_samples 19 [1046712, 70047] processed_samples 6300 unjoint_samples 6300 joint_samples 17 [558406, 1041805] processed_samples 6300 unjoint_samples 6300 joint_samples 17 [558406, 1041805] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [464371, 1047110] processed_samples 6300 unjoint_samples 6300 joint_samples 18 [464371, 1047110] processed_samples 6301 unjoint_samples 6300 joint_samples 17 [255653, 1046661] processed_samples 6301 unjoint_samples 6300 joint_samples 18 [717446, 1005743] processed_samples 6301 unjoint_samples 6300 joint_samples 17 [255653, 1046661] processed_samples 6301 unjoint_samples 6300 joint_samples 18 [717446, 1005743] processed_samples 6301 unjoint_samples 6300 joint_samples 17 [1046010, 1032738] processed_samples 6301 unjoint_samples 6300 joint_samples 17 [1046010, 1032738] [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure processed_samples 6400 unjoint_samples 6400 joint_samples 19 [1046712, 288225] processed_samples 6400 unjoint_samples 6400 joint_samples 19 [1046712, 288225] processed_samples 6400 unjoint_samples 6400 joint_samples 17 [801672, 1041805] processed_samples 6400 unjoint_samples 6400 joint_samples 17 [801672, 1041805] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [745415, 1047110] processed_samples 6400 unjoint_samples 6400 joint_samples 18 [910857, 1033978] processed_samples 6401 unjoint_samples 6400 joint_samples 18 [254794, 1045882] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure processed_samples 6400 unjoint_samples 6400 joint_samples 18 [745415, 1047110] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure processed_samples 6400 unjoint_samples 6400 joint_samples 18 [910857, 1033978] processed_samples 6401 unjoint_samples 6400 joint_samples 18 [254794, 1045882] [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure processed_samples 6401 unjoint_samples 6400 joint_samples 17 [530899, 1046661] processed_samples 6401 unjoint_samples 6400 joint_samples 17 [530899, 1046661] processed_samples 6401 unjoint_samples 6400 joint_samples 18 [1044115, 1043912] processed_samples 6400 unjoint_samples 6400 joint_samples 17 [938560, 1045044] processed_samples 6401 unjoint_samples 6400 joint_samples 18 [1044115, 1043912] processed_samples 6400 unjoint_samples 6400 joint_samples 17 [938560, 1045044] [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea4324a80] mmco: unref short failure [h264 @ 0x561ea4324a80] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure processed_samples 6500 unjoint_samples 6500 joint_samples 19 [164720, 1046769] processed_samples 6500 unjoint_samples 6500 joint_samples 19 [164720, 1046769] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1013634, 242535] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1013634, 242535] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [78921, 1047642] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [78921, 1047642] processed_samples 6501 unjoint_samples 6500 joint_samples 18 [567749, 1045882] processed_samples 6500 unjoint_samples 6500 joint_samples 19 [1046712, 689768] processed_samples 6500 unjoint_samples 6500 joint_samples 19 [1046712, 689768] processed_samples 6501 unjoint_samples 6500 joint_samples 18 [567749, 1045882] processed_samples 6501 unjoint_samples 6500 joint_samples 19 [354419, 1047015] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1039590, 1047110] processed_samples 6501 unjoint_samples 6500 joint_samples 17 [875628, 1046661] processed_samples 6501 unjoint_samples 6500 joint_samples 19 [354419, 1047015] processed_samples 6501 unjoint_samples 6500 joint_samples 17 [875628, 1046661] processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1039590, 1047110] [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea15e3240] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1013634, 592050] processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1013634, 592050] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [390599, 1046769] processed_samples 6600 unjoint_samples 6600 joint_samples 18 [400605, 1047642] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [390599, 1046769] processed_samples 6600 unjoint_samples 6600 joint_samples 18 [400605, 1047642] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [1041845, 273911] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [1041845, 273911] processed_samples 6601 unjoint_samples 6600 joint_samples 18 [1035272, 294085] processed_samples 6601 unjoint_samples 6600 joint_samples 18 [1035272, 294085] processed_samples 6601 unjoint_samples 6600 joint_samples 19 [725730, 1047015] processed_samples 6601 unjoint_samples 6600 joint_samples 19 [725730, 1047015] processed_samples 6601 unjoint_samples 6600 joint_samples 18 [824778, 1045882] processed_samples 6601 unjoint_samples 6600 joint_samples 18 [824778, 1045882] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [1046712, 980754] processed_samples 6600 unjoint_samples 6600 joint_samples 19 [1046712, 980754] [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure processed_samples 6700 unjoint_samples 6700 joint_samples 18 [1013634, 836103] [h264 @ 0x561ea1b67a00] mmco: unref short failure processed_samples 6700 unjoint_samples 6700 joint_samples 18 [1013634, 836103] processed_samples 6700 unjoint_samples 6700 joint_samples 18 [694296, 1047642] processed_samples 6700 unjoint_samples 6700 joint_samples 18 [694296, 1047642] [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure processed_samples 6701 unjoint_samples 6700 joint_samples 19 [124419, 1047217] processed_samples 6700 unjoint_samples 6700 joint_samples 20 [1046712, 415035] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [1041845, 666510] processed_samples 6701 unjoint_samples 6700 joint_samples 19 [124419, 1047217] processed_samples 6700 unjoint_samples 6700 joint_samples 20 [1046712, 415035] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [692722, 1046769] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [1041845, 666510] processed_samples 6700 unjoint_samples 6700 joint_samples 19 [692722, 1046769] processed_samples 6701 unjoint_samples 6700 joint_samples 18 [1035272, 627742] processed_samples 6701 unjoint_samples 6700 joint_samples 18 [1035272, 627742] processed_samples 6701 unjoint_samples 6700 joint_samples 19 [987247, 1047015] processed_samples 6701 unjoint_samples 6700 joint_samples 19 [987247, 1047015] [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure processed_samples 6800 unjoint_samples 6800 joint_samples 19 [62297, 1043654] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [62297, 1043654] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure processed_samples 6801 unjoint_samples 6800 joint_samples 18 [1035272, 955290] processed_samples 6800 unjoint_samples 6800 joint_samples 20 [1046712, 709264] processed_samples 6800 unjoint_samples 6800 joint_samples 20 [1046712, 709264] processed_samples 6801 unjoint_samples 6800 joint_samples 18 [1035272, 955290] processed_samples 6801 unjoint_samples 6800 joint_samples 20 [1019486, 351740] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [941697, 1046769] processed_samples 6801 unjoint_samples 6800 joint_samples 19 [466386, 1047217] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [941697, 1046769] processed_samples 6801 unjoint_samples 6800 joint_samples 19 [466386, 1047217] processed_samples 6801 unjoint_samples 6800 joint_samples 20 [1019486, 351740] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [1041845, 969894] processed_samples 6800 unjoint_samples 6800 joint_samples 19 [1041845, 969894] processed_samples 6800 unjoint_samples 6800 joint_samples 18 [1025058, 1047642] processed_samples 6800 unjoint_samples 6800 joint_samples 18 [1025058, 1047642] [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1039423, 231505] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1039423, 231505] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1046548, 227527] processed_samples 6900 unjoint_samples 6900 joint_samples 20 [1046548, 227527] processed_samples 6900 unjoint_samples 6900 joint_samples 19 [506914, 1043654] processed_samples 6900 unjoint_samples 6900 joint_samples 19 [506914, 1043654] processed_samples 6900 unjoint_samples 6900 joint_samples 19 [1046409, 353465] processed_samples 6900 unjoint_samples 6900 joint_samples 19 [1046409, 353465] processed_samples 6900 unjoint_samples 6900 joint_samples 21 [28980, 1045164] processed_samples 6900 unjoint_samples 6900 joint_samples 21 [28980, 1045164] processed_samples 6901 unjoint_samples 6900 joint_samples 19 [190950, 1035970] processed_samples 6901 unjoint_samples 6900 joint_samples 19 [190950, 1035970] [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure processed_samples 6901 unjoint_samples 6900 joint_samples 20 [1019486, 876491] processed_samples 6901 unjoint_samples 6900 joint_samples 20 [1019486, 876491] processed_samples 6901 unjoint_samples 6900 joint_samples 19 [722426, 1047217] processed_samples 6901 unjoint_samples 6900 joint_samples 19 [722426, 1047217] [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1046548, 513157] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1046548, 513157] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1039423, 622710] processed_samples 7000 unjoint_samples 7000 joint_samples 21 [321304, 1045164] processed_samples 7000 unjoint_samples 7000 joint_samples 21 [321304, 1045164] processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1039423, 622710] processed_samples 7001 unjoint_samples 7000 joint_samples 21 [146725, 1021002] processed_samples 7000 unjoint_samples 7000 joint_samples 19 [938197, 1043654] processed_samples 7001 unjoint_samples 7000 joint_samples 21 [146725, 1021002] [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure processed_samples 7000 unjoint_samples 7000 joint_samples 19 [938197, 1043654] processed_samples 7000 unjoint_samples 7000 joint_samples 19 [1046409, 703673] processed_samples 7000 unjoint_samples 7000 joint_samples 19 [1046409, 703673] processed_samples 7001 unjoint_samples 7000 joint_samples 19 [518376, 1035970] processed_samples 7001 unjoint_samples 7000 joint_samples 19 [518376, 1035970] processed_samples 7001 unjoint_samples 7000 joint_samples 19 [1015327, 1047217] processed_samples 7001 unjoint_samples 7000 joint_samples 19 [1015327, 1047217] [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1018454, 235557] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1018454, 235557] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [99197, 1045131] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [99197, 1045131] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1046548, 773472] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1046548, 773472] processed_samples 7100 unjoint_samples 7100 joint_samples 21 [608218, 1045164] processed_samples 7100 unjoint_samples 7100 joint_samples 21 [608218, 1045164] processed_samples 7101 unjoint_samples 7100 joint_samples 21 [439186, 1021002] processed_samples 7101 unjoint_samples 7100 joint_samples 21 [439186, 1021002] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1039423, 909376] processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1039423, 909376] processed_samples 7101 unjoint_samples 7100 joint_samples 19 [783540, 1035970] processed_samples 7101 unjoint_samples 7100 joint_samples 19 [783540, 1035970] [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 7101 unjoint_samples 7100 joint_samples 20 [1018643, 391475] processed_samples 7101 unjoint_samples 7100 joint_samples 20 [1018643, 391475] [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1018454, 616352] processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1018454, 616352] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [191313, 1040067] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [191313, 1040067] processed_samples 7200 unjoint_samples 7200 joint_samples 20 [454980, 1045131] processed_samples 7200 unjoint_samples 7200 joint_samples 20 [454980, 1045131] processed_samples 7201 unjoint_samples 7200 joint_samples 21 [761919, 1021002] processed_samples 7201 unjoint_samples 7200 joint_samples 21 [761919, 1021002] processed_samples 7201 unjoint_samples 7200 joint_samples 20 [1044277, 155078] processed_samples 7201 unjoint_samples 7200 joint_samples 20 [1044277, 155078] processed_samples 7201 unjoint_samples 7200 joint_samples 20 [1018643, 788274] processed_samples 7201 unjoint_samples 7200 joint_samples 20 [1018643, 788274] processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1046548, 995621] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [899222, 1045164] processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1046548, 995621] processed_samples 7200 unjoint_samples 7200 joint_samples 21 [899222, 1045164] [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure processed_samples 7300 unjoint_samples 7300 joint_samples 21 [432467, 1040067] [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure processed_samples 7300 unjoint_samples 7300 joint_samples 21 [432467, 1040067] processed_samples 7300 unjoint_samples 7300 joint_samples 20 [724935, 1045131] processed_samples 7300 unjoint_samples 7300 joint_samples 20 [724935, 1045131] processed_samples 7300 unjoint_samples 7300 joint_samples 22 [1046909, 166165] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [1046548, 352424] processed_samples 7300 unjoint_samples 7300 joint_samples 22 [1046909, 166165] processed_samples 7300 unjoint_samples 7300 joint_samples 21 [1046548, 352424] processed_samples 7300 unjoint_samples 7300 joint_samples 20 [1018454, 987352] processed_samples 7300 unjoint_samples 7300 joint_samples 20 [1018454, 987352] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure processed_samples 7301 unjoint_samples 7300 joint_samples 20 [1044277, 544085] processed_samples 7301 unjoint_samples 7300 joint_samples 20 [1044277, 544085] processed_samples 7301 unjoint_samples 7300 joint_samples 20 [1020898, 1020713] processed_samples 7301 unjoint_samples 7300 joint_samples 22 [1031083, 245699] processed_samples 7301 unjoint_samples 7300 joint_samples 22 [1031083, 245699] processed_samples 7301 unjoint_samples 7300 joint_samples 20 [1020898, 1020713] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure processed_samples 7400 unjoint_samples 7400 joint_samples 21 [272363, 1034291] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [272363, 1034291] [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure processed_samples 7401 unjoint_samples 7400 joint_samples 21 [1046457, 241358] processed_samples 7401 unjoint_samples 7400 joint_samples 21 [1046457, 241358] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [729815, 1040067] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [729815, 1040067] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1046548, 703380] processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1046548, 703380] processed_samples 7400 unjoint_samples 7400 joint_samples 22 [1046909, 508473] processed_samples 7401 unjoint_samples 7400 joint_samples 20 [1044277, 834742] processed_samples 7400 unjoint_samples 7400 joint_samples 22 [1046909, 508473] processed_samples 7401 unjoint_samples 7400 joint_samples 20 [1044277, 834742] processed_samples 7401 unjoint_samples 7400 joint_samples 22 [1031083, 541285] processed_samples 7400 unjoint_samples 7400 joint_samples 20 [976844, 1045131] processed_samples 7401 unjoint_samples 7400 joint_samples 22 [1031083, 541285] processed_samples 7400 unjoint_samples 7400 joint_samples 20 [976844, 1045131] [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure processed_samples 7500 unjoint_samples 7500 joint_samples 21 [243630, 1047249] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [578947, 1034291] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [243630, 1047249] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [1046909, 795409] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [1046548, 1016041] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [990168, 1040067] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [578947, 1034291] processed_samples 7501 unjoint_samples 7500 joint_samples 21 [1046457, 540875] processed_samples 7500 unjoint_samples 7500 joint_samples 22 [1046909, 795409] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [1046548, 1016041] processed_samples 7501 unjoint_samples 7500 joint_samples 21 [287193, 1046958] processed_samples 7500 unjoint_samples 7500 joint_samples 21 [990168, 1040067] processed_samples 7501 unjoint_samples 7500 joint_samples 21 [1046457, 540875] processed_samples 7501 unjoint_samples 7500 joint_samples 21 [287193, 1046958] processed_samples 7501 unjoint_samples 7500 joint_samples 22 [1031083, 859130] processed_samples 7501 unjoint_samples 7500 joint_samples 22 [1031083, 859130] [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure processed_samples 7600 unjoint_samples 7600 joint_samples 23 [71632, 1034231] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1047314, 293622] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1047314, 293622] processed_samples 7600 unjoint_samples 7600 joint_samples 23 [71632, 1034231] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [295598, 1045719] processed_samples 7600 unjoint_samples 7600 joint_samples 22 [295598, 1045719] processed_samples 7600 unjoint_samples 7600 joint_samples 21 [556148, 1047249] processed_samples 7600 unjoint_samples 7600 joint_samples 21 [556148, 1047249] processed_samples 7601 unjoint_samples 7600 joint_samples 21 [1046457, 824211] processed_samples 7601 unjoint_samples 7600 joint_samples 21 [1046457, 824211] processed_samples 7601 unjoint_samples 7600 joint_samples 21 [639356, 1046958] processed_samples 7601 unjoint_samples 7600 joint_samples 21 [639356, 1046958] processed_samples 7600 unjoint_samples 7600 joint_samples 21 [956229, 1034291] processed_samples 7600 unjoint_samples 7600 joint_samples 21 [956229, 1034291] processed_samples 7601 unjoint_samples 7600 joint_samples 23 [302736, 1017497] processed_samples 7601 unjoint_samples 7600 joint_samples 23 [302736, 1017497] [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure processed_samples 7700 unjoint_samples 7700 joint_samples 22 [160710, 1047522] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1047314, 540569] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1047314, 540569] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [160710, 1047522] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [645230, 1045719] processed_samples 7700 unjoint_samples 7700 joint_samples 22 [645230, 1045719] processed_samples 7700 unjoint_samples 7700 joint_samples 23 [368193, 1034231] processed_samples 7700 unjoint_samples 7700 joint_samples 23 [368193, 1034231] processed_samples 7701 unjoint_samples 7700 joint_samples 22 [1046457, 198026] processed_samples 7700 unjoint_samples 7700 joint_samples 21 [799296, 1047249] processed_samples 7700 unjoint_samples 7700 joint_samples 21 [799296, 1047249] processed_samples 7701 unjoint_samples 7700 joint_samples 22 [1046457, 198026] processed_samples 7701 unjoint_samples 7700 joint_samples 23 [658645, 1017497] processed_samples 7701 unjoint_samples 7700 joint_samples 23 [658645, 1017497] processed_samples 7701 unjoint_samples 7700 joint_samples 21 [897563, 1046958] processed_samples 7701 unjoint_samples 7700 joint_samples 21 [897563, 1046958] [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure processed_samples 7800 unjoint_samples 7800 joint_samples 22 [416185, 1047522] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1019770, 80328] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [416185, 1047522] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1019770, 80328] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [913319, 1045719] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [913319, 1045719] processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1047314, 964606] processed_samples 7801 unjoint_samples 7800 joint_samples 22 [1046457, 547435] processed_samples 7801 unjoint_samples 7800 joint_samples 22 [1046457, 547435] [h264 @ 0x561ea429e380] mmco: unref short failure processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1047314, 964606] [h264 @ 0x55e6cea6c680] mmco: unref short failure processed_samples 7800 unjoint_samples 7800 joint_samples 23 [589921, 1034231] processed_samples 7800 unjoint_samples 7800 joint_samples 23 [589921, 1034231] processed_samples 7801 unjoint_samples 7800 joint_samples 22 [167355, 1046958] processed_samples 7801 unjoint_samples 7800 joint_samples 22 [167355, 1046958] processed_samples 7801 unjoint_samples 7800 joint_samples 23 [1040931, 1041500] processed_samples 7801 unjoint_samples 7800 joint_samples 23 [1040931, 1041500] [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1034041, 199953] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [287406, 1045964] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1034041, 199953] processed_samples 7901 unjoint_samples 7900 joint_samples 24 [388400, 1047035] processed_samples 7901 unjoint_samples 7900 joint_samples 24 [388400, 1047035] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [287406, 1045964] processed_samples 7900 unjoint_samples 7900 joint_samples 22 [1019770, 325786] processed_samples 7900 unjoint_samples 7900 joint_samples 22 [1019770, 325786] processed_samples 7900 unjoint_samples 7900 joint_samples 22 [712714, 1047522] processed_samples 7900 unjoint_samples 7900 joint_samples 22 [712714, 1047522] processed_samples 7901 unjoint_samples 7900 joint_samples 22 [1046457, 878146] processed_samples 7901 unjoint_samples 7900 joint_samples 22 [1046457, 878146] processed_samples 7901 unjoint_samples 7900 joint_samples 22 [461374, 1046958] processed_samples 7901 unjoint_samples 7900 joint_samples 22 [461374, 1046958] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [956306, 1034231] processed_samples 7900 unjoint_samples 7900 joint_samples 23 [956306, 1034231] [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure processed_samples 8000 unjoint_samples 8000 joint_samples 23 [666501, 1045964] processed_samples 8000 unjoint_samples 8000 joint_samples 24 [256495, 1034231] processed_samples 8000 unjoint_samples 8000 joint_samples 22 [1019770, 614705] processed_samples 8000 unjoint_samples 8000 joint_samples 22 [1019770, 614705] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1034041, 505485] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [666501, 1045964] processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1034041, 505485] processed_samples 8000 unjoint_samples 8000 joint_samples 24 [256495, 1034231] processed_samples 8001 unjoint_samples 8000 joint_samples 23 [1046457, 100795] processed_samples 8001 unjoint_samples 8000 joint_samples 23 [1046457, 100795] processed_samples 8001 unjoint_samples 8000 joint_samples 24 [728294, 1047035] processed_samples 8001 unjoint_samples 8000 joint_samples 24 [728294, 1047035] processed_samples 8001 unjoint_samples 8000 joint_samples 22 [855930, 1046958] processed_samples 8001 unjoint_samples 8000 joint_samples 22 [855930, 1046958] processed_samples 8000 unjoint_samples 8000 joint_samples 22 [1047457, 1047522] [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure processed_samples 8000 unjoint_samples 8000 joint_samples 22 [1047457, 1047522] [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1047457, 422656] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1047457, 422656] processed_samples 8101 unjoint_samples 8100 joint_samples 23 [1045785, 177255] processed_samples 8100 unjoint_samples 8100 joint_samples 24 [583505, 1034231] processed_samples 8100 unjoint_samples 8100 joint_samples 24 [583505, 1034231] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1034041, 810085] processed_samples 8101 unjoint_samples 8100 joint_samples 23 [1046457, 383899] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1044253, 1045964] processed_samples 8100 unjoint_samples 8100 joint_samples 22 [1019770, 1018736] processed_samples 8101 unjoint_samples 8100 joint_samples 23 [1045785, 177255] processed_samples 8100 unjoint_samples 8100 joint_samples 22 [1019770, 1018736] [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1034041, 810085] processed_samples 8101 unjoint_samples 8100 joint_samples 24 [1012929, 1047035] processed_samples 8101 unjoint_samples 8100 joint_samples 24 [1012929, 1047035] processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1044253, 1045964] processed_samples 8101 unjoint_samples 8100 joint_samples 23 [1046457, 383899] [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1045814, 274137] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1045814, 274137] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1046512, 329526] processed_samples 8200 unjoint_samples 8200 joint_samples 23 [345420, 1021220] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1046512, 329526] processed_samples 8200 unjoint_samples 8200 joint_samples 23 [345420, 1021220] processed_samples 8201 unjoint_samples 8200 joint_samples 25 [1046662, 397329] processed_samples 8201 unjoint_samples 8200 joint_samples 25 [1046662, 397329] processed_samples 8201 unjoint_samples 8200 joint_samples 23 [1045785, 451316] processed_samples 8201 unjoint_samples 8200 joint_samples 23 [1045785, 451316] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure processed_samples 8200 unjoint_samples 8200 joint_samples 23 [1047457, 771059] processed_samples 8200 unjoint_samples 8200 joint_samples 23 [1047457, 771059] processed_samples 8201 unjoint_samples 8200 joint_samples 23 [1046457, 632438] processed_samples 8201 unjoint_samples 8200 joint_samples 23 [1046457, 632438] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [846593, 1034231] processed_samples 8200 unjoint_samples 8200 joint_samples 24 [846593, 1034231] [h264 @ 0x55e6cdfe92c0] Missing reference picture, default is 65524 [h264 @ 0x561ea3ffc480] Missing reference picture, default is 65524 [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6cec12f80] mmco: unref short failure processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1046512, 571340] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1046512, 571340] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [31564, 1041084] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [31564, 1041084] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1045814, 578698] processed_samples 8300 unjoint_samples 8300 joint_samples 25 [55918, 1036597] processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1045814, 578698] processed_samples 8300 unjoint_samples 8300 joint_samples 25 [55918, 1036597] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure processed_samples 8300 unjoint_samples 8300 joint_samples 23 [628499, 1021220] processed_samples 8300 unjoint_samples 8300 joint_samples 23 [628499, 1021220] processed_samples 8301 unjoint_samples 8300 joint_samples 23 [1045785, 825013] processed_samples 8301 unjoint_samples 8300 joint_samples 23 [1045785, 825013] processed_samples 8301 unjoint_samples 8300 joint_samples 25 [1046662, 680423] processed_samples 8301 unjoint_samples 8300 joint_samples 25 [1046662, 680423] processed_samples 8301 unjoint_samples 8300 joint_samples 23 [1046457, 959418] processed_samples 8301 unjoint_samples 8300 joint_samples 23 [1046457, 959418] [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure processed_samples 8400 unjoint_samples 8400 joint_samples 24 [416844, 1041084] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [416844, 1041084] processed_samples 8400 unjoint_samples 8400 joint_samples 25 [328624, 1036597] processed_samples 8400 unjoint_samples 8400 joint_samples 25 [328624, 1036597] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1046512, 974052] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1046512, 974052] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1045814, 836264] processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1045814, 836264] processed_samples 8401 unjoint_samples 8400 joint_samples 24 [1046457, 183643] processed_samples 8401 unjoint_samples 8400 joint_samples 24 [1046457, 183643] processed_samples 8401 unjoint_samples 8400 joint_samples 24 [1045785, 153738] processed_samples 8401 unjoint_samples 8400 joint_samples 24 [1045785, 153738] [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure processed_samples 8400 unjoint_samples 8400 joint_samples 23 [938815, 1021220] processed_samples 8400 unjoint_samples 8400 joint_samples 23 [938815, 1021220] processed_samples 8401 unjoint_samples 8400 joint_samples 25 [1046662, 946239] processed_samples 8401 unjoint_samples 8400 joint_samples 25 [1046662, 946239] [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure processed_samples 8500 unjoint_samples 8500 joint_samples 25 [330764, 1041783] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [330764, 1041783] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1045814, 96687] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1045814, 96687] processed_samples 8500 unjoint_samples 8500 joint_samples 24 [651007, 1041084] processed_samples 8500 unjoint_samples 8500 joint_samples 24 [651007, 1041084] processed_samples 8500 unjoint_samples 8500 joint_samples 24 [1047746, 156578] processed_samples 8501 unjoint_samples 8500 joint_samples 26 [1046662, 179970] processed_samples 8500 unjoint_samples 8500 joint_samples 24 [1047746, 156578] processed_samples 8501 unjoint_samples 8500 joint_samples 26 [1046662, 179970] processed_samples 8501 unjoint_samples 8500 joint_samples 24 [1046457, 509857] processed_samples 8501 unjoint_samples 8500 joint_samples 24 [1046457, 509857] processed_samples 8501 unjoint_samples 8500 joint_samples 24 [1045785, 436838] processed_samples 8501 unjoint_samples 8500 joint_samples 24 [1045785, 436838] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [634884, 1036597] processed_samples 8500 unjoint_samples 8500 joint_samples 25 [634884, 1036597] [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure processed_samples 8600 unjoint_samples 8600 joint_samples 24 [1047746, 536584] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1045814, 389313] processed_samples 8600 unjoint_samples 8600 joint_samples 24 [900470, 1041084] [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure processed_samples 8601 unjoint_samples 8600 joint_samples 24 [1046457, 795304] processed_samples 8600 unjoint_samples 8600 joint_samples 24 [1047746, 536584] processed_samples 8601 unjoint_samples 8600 joint_samples 26 [1046662, 494228] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [640424, 1041783] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1045814, 389313] processed_samples 8600 unjoint_samples 8600 joint_samples 24 [900470, 1041084] processed_samples 8602 unjoint_samples 8600 joint_samples 24 [1045785, 723576] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [971971, 1036597] processed_samples 8601 unjoint_samples 8600 joint_samples 24 [1046457, 795304] processed_samples 8601 unjoint_samples 8600 joint_samples 26 [1046662, 494228] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [640424, 1041783] processed_samples 8600 unjoint_samples 8600 joint_samples 25 [971971, 1036597] processed_samples 8602 unjoint_samples 8600 joint_samples 24 [1045785, 723576] [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1014606, 272258] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1014606, 272258] processed_samples 8701 unjoint_samples 8700 joint_samples 26 [1046662, 735436] processed_samples 8701 unjoint_samples 8700 joint_samples 26 [1046662, 735436] processed_samples 8700 unjoint_samples 8700 joint_samples 26 [1047363, 211328] processed_samples 8700 unjoint_samples 8700 joint_samples 24 [1047746, 798460] processed_samples 8701 unjoint_samples 8700 joint_samples 25 [129607, 1046018] processed_samples 8700 unjoint_samples 8700 joint_samples 26 [1047363, 211328] processed_samples 8701 unjoint_samples 8700 joint_samples 25 [129607, 1046018] processed_samples 8700 unjoint_samples 8700 joint_samples 24 [1047746, 798460] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1045814, 772300] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1045814, 772300] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [920725, 1041783] processed_samples 8700 unjoint_samples 8700 joint_samples 25 [920725, 1041783] processed_samples 8702 unjoint_samples 8700 joint_samples 25 [96897, 1036519] processed_samples 8702 unjoint_samples 8700 joint_samples 25 [96897, 1036519] [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 8800 unjoint_samples 8800 joint_samples 26 [64352, 1047597] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [64352, 1047597] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1027396, 218123] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1027396, 218123] processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1014606, 546942] processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1014606, 546942] processed_samples 8800 unjoint_samples 8800 joint_samples 25 [164687, 1035878] processed_samples 8800 unjoint_samples 8800 joint_samples 25 [164687, 1035878] processed_samples 8801 unjoint_samples 8800 joint_samples 25 [410713, 1046018] processed_samples 8801 unjoint_samples 8800 joint_samples 25 [410713, 1046018] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1047363, 459092] processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1047363, 459092] processed_samples 8801 unjoint_samples 8800 joint_samples 26 [1046662, 986744] processed_samples 8801 unjoint_samples 8800 joint_samples 26 [1046662, 986744] processed_samples 8802 unjoint_samples 8800 joint_samples 25 [530279, 1036519] processed_samples 8802 unjoint_samples 8800 joint_samples 25 [530279, 1036519] [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure processed_samples 8900 unjoint_samples 8900 joint_samples 26 [455199, 1047597] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [455199, 1047597] processed_samples 8900 unjoint_samples 8900 joint_samples 25 [469141, 1035878] processed_samples 8900 unjoint_samples 8900 joint_samples 25 [469141, 1035878] processed_samples 8900 unjoint_samples 8900 joint_samples 25 [1014606, 926356] processed_samples 8900 unjoint_samples 8900 joint_samples 25 [1014606, 926356] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1027396, 584794] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1027396, 584794] processed_samples 8901 unjoint_samples 8900 joint_samples 25 [763522, 1046018] processed_samples 8901 unjoint_samples 8900 joint_samples 25 [763522, 1046018] processed_samples 8901 unjoint_samples 8900 joint_samples 27 [1046662, 229810] processed_samples 8901 unjoint_samples 8900 joint_samples 27 [1046662, 229810] [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1047363, 759970] processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1047363, 759970] processed_samples 8902 unjoint_samples 8900 joint_samples 25 [853144, 1036519] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure processed_samples 8902 unjoint_samples 8900 joint_samples 25 [853144, 1036519] [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4c60e40] Missing reference picture, default is 65530 [h264 @ 0x561ea4c60e40] Missing reference picture, default is 65530 [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] Missing reference picture, default is 65530 [h264 @ 0x561ea4c60e40] Missing reference picture, default is 65530 [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d16cf440] Missing reference picture, default is 65530 [h264 @ 0x55e6d16cf440] Missing reference picture, default is 65530 [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] Missing reference picture, default is 65530 [h264 @ 0x55e6d16cf440] Missing reference picture, default is 65530 [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure processed_samples 9000 unjoint_samples 9000 joint_samples 26 [188693, 1044578] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [188693, 1044578] processed_samples 9000 unjoint_samples 9000 joint_samples 25 [872370, 1035878] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [738810, 1047597] processed_samples 9000 unjoint_samples 9000 joint_samples 25 [872370, 1035878] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [738810, 1047597] processed_samples 9001 unjoint_samples 9000 joint_samples 26 [1043369, 141088] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1027396, 843430] processed_samples 9001 unjoint_samples 9000 joint_samples 26 [1043369, 141088] processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1027396, 843430] processed_samples 9001 unjoint_samples 9000 joint_samples 27 [1046662, 528020] processed_samples 9001 unjoint_samples 9000 joint_samples 27 [1046662, 528020] processed_samples 9000 unjoint_samples 9000 joint_samples 27 [1047363, 5751] processed_samples 9000 unjoint_samples 9000 joint_samples 27 [1047363, 5751] [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure processed_samples 9002 unjoint_samples 9000 joint_samples 26 [1047205, 64857] processed_samples 9002 unjoint_samples 9000 joint_samples 26 [1047205, 64857] [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure processed_samples 9100 unjoint_samples 9100 joint_samples 27 [1037026, 37312] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [1037026, 37312] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [1038776, 147251] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [151048, 1035878] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [1038776, 147251] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [151048, 1035878] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [1047363, 274106] processed_samples 9100 unjoint_samples 9100 joint_samples 27 [1047363, 274106] processed_samples 9101 unjoint_samples 9100 joint_samples 26 [1043369, 411646] processed_samples 9101 unjoint_samples 9100 joint_samples 26 [1043369, 411646] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [650563, 1044578] processed_samples 9100 unjoint_samples 9100 joint_samples 26 [650563, 1044578] processed_samples 9102 unjoint_samples 9100 joint_samples 26 [1047205, 310728] processed_samples 9102 unjoint_samples 9100 joint_samples 26 [1047205, 310728] processed_samples 9101 unjoint_samples 9100 joint_samples 27 [1046662, 809361] processed_samples 9101 unjoint_samples 9100 joint_samples 27 [1046662, 809361] [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1038776, 499387] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1038776, 499387] processed_samples 9200 unjoint_samples 9200 joint_samples 26 [405554, 1035878] processed_samples 9200 unjoint_samples 9200 joint_samples 26 [405554, 1035878] processed_samples 9200 unjoint_samples 9200 joint_samples 26 [988588, 1044578] processed_samples 9200 unjoint_samples 9200 joint_samples 26 [988588, 1044578] [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1047363, 555380] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1047363, 555380] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure processed_samples 9201 unjoint_samples 9200 joint_samples 26 [1043369, 663209] processed_samples 9201 unjoint_samples 9200 joint_samples 26 [1043369, 663209] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1037026, 336489] processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1037026, 336489] processed_samples 9201 unjoint_samples 9200 joint_samples 28 [1046662, 69584] processed_samples 9201 unjoint_samples 9200 joint_samples 28 [1046662, 69584] [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure processed_samples 9202 unjoint_samples 9200 joint_samples 26 [1047205, 607968] processed_samples 9202 unjoint_samples 9200 joint_samples 26 [1047205, 607968] [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1038776, 779890] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [188110, 1044578] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1038776, 779890] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [188110, 1044578] processed_samples 9300 unjoint_samples 9300 joint_samples 26 [740910, 1035878] processed_samples 9300 unjoint_samples 9300 joint_samples 26 [740910, 1035878] processed_samples 9301 unjoint_samples 9300 joint_samples 28 [1046662, 328667] processed_samples 9301 unjoint_samples 9300 joint_samples 28 [1046662, 328667] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1037026, 606700] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1037026, 606700] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1047363, 854621] processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1047363, 854621] processed_samples 9302 unjoint_samples 9300 joint_samples 26 [1047205, 995186] processed_samples 9301 unjoint_samples 9300 joint_samples 26 [1043369, 1045116] processed_samples 9301 unjoint_samples 9300 joint_samples 26 [1043369, 1045116] processed_samples 9302 unjoint_samples 9300 joint_samples 26 [1047205, 995186] [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure processed_samples 9400 unjoint_samples 9400 joint_samples 28 [117276, 993481] processed_samples 9400 unjoint_samples 9400 joint_samples 28 [117276, 993481] processed_samples 9400 unjoint_samples 9400 joint_samples 28 [1047363, 222186] processed_samples 9400 unjoint_samples 9400 joint_samples 28 [1047363, 222186] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [410298, 1044578] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [410298, 1044578] processed_samples 9401 unjoint_samples 9400 joint_samples 27 [1048370, 335810] processed_samples 9401 unjoint_samples 9400 joint_samples 27 [1048370, 335810] [h264 @ 0x561ea27ac400] mmco: unref short failure processed_samples 9400 unjoint_samples 9400 joint_samples 27 [23008, 1040039] [h264 @ 0x55e6ce443040] mmco: unref short failure processed_samples 9400 unjoint_samples 9400 joint_samples 27 [23008, 1040039] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1037026, 922388] processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1037026, 922388] processed_samples 9401 unjoint_samples 9400 joint_samples 28 [1046662, 801173] processed_samples 9401 unjoint_samples 9400 joint_samples 28 [1046662, 801173] processed_samples 9402 unjoint_samples 9400 joint_samples 27 [1047205, 394014] processed_samples 9402 unjoint_samples 9400 joint_samples 27 [1047205, 394014] [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure processed_samples 9500 unjoint_samples 9500 joint_samples 28 [1037026, 171306] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [1037026, 171306] processed_samples 9500 unjoint_samples 9500 joint_samples 27 [324921, 1040039] [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure processed_samples 9500 unjoint_samples 9500 joint_samples 27 [324921, 1040039] [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure processed_samples 9500 unjoint_samples 9500 joint_samples 28 [370896, 993481] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [370896, 993481] processed_samples 9500 unjoint_samples 9500 joint_samples 27 [708691, 1044578] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [1047363, 516093] processed_samples 9500 unjoint_samples 9500 joint_samples 27 [708691, 1044578] processed_samples 9500 unjoint_samples 9500 joint_samples 28 [1047363, 516093] processed_samples 9501 unjoint_samples 9500 joint_samples 27 [1048370, 737129] processed_samples 9501 unjoint_samples 9500 joint_samples 27 [1048370, 737129] processed_samples 9501 unjoint_samples 9500 joint_samples 28 [1046662, 1048038] processed_samples 9501 unjoint_samples 9500 joint_samples 28 [1046662, 1048038] processed_samples 9502 unjoint_samples 9500 joint_samples 27 [1047205, 820663] processed_samples 9502 unjoint_samples 9500 joint_samples 27 [1047205, 820663] [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure processed_samples 9600 unjoint_samples 9600 joint_samples 27 [980992, 1044578] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1047363, 805150] processed_samples 9601 unjoint_samples 9600 joint_samples 28 [114418, 1035681] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1037026, 470693] processed_samples 9600 unjoint_samples 9600 joint_samples 27 [671312, 1040039] processed_samples 9601 unjoint_samples 9600 joint_samples 29 [1047214, 341164] processed_samples 9600 unjoint_samples 9600 joint_samples 27 [980992, 1044578] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1047363, 805150] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [669450, 993481] processed_samples 9601 unjoint_samples 9600 joint_samples 28 [114418, 1035681] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1037026, 470693] processed_samples 9600 unjoint_samples 9600 joint_samples 27 [671312, 1040039] processed_samples 9601 unjoint_samples 9600 joint_samples 29 [1047214, 341164] processed_samples 9600 unjoint_samples 9600 joint_samples 28 [669450, 993481] processed_samples 9602 unjoint_samples 9600 joint_samples 28 [11547, 1041374] processed_samples 9602 unjoint_samples 9600 joint_samples 28 [11547, 1041374] [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure processed_samples 9700 unjoint_samples 9700 joint_samples 28 [196315, 1046850] [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure processed_samples 9700 unjoint_samples 9700 joint_samples 28 [196315, 1046850] processed_samples 9700 unjoint_samples 9700 joint_samples 29 [1048080, 60965] processed_samples 9701 unjoint_samples 9700 joint_samples 28 [393231, 1035681] processed_samples 9700 unjoint_samples 9700 joint_samples 29 [1048080, 60965] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [1037026, 716694] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [962181, 993481] processed_samples 9701 unjoint_samples 9700 joint_samples 28 [393231, 1035681] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [1037026, 716694] processed_samples 9700 unjoint_samples 9700 joint_samples 27 [1021389, 1040039] processed_samples 9701 unjoint_samples 9700 joint_samples 29 [1047214, 654332] processed_samples 9702 unjoint_samples 9700 joint_samples 28 [314455, 1041374] processed_samples 9700 unjoint_samples 9700 joint_samples 27 [1021389, 1040039] processed_samples 9700 unjoint_samples 9700 joint_samples 28 [962181, 993481] processed_samples 9701 unjoint_samples 9700 joint_samples 29 [1047214, 654332] processed_samples 9702 unjoint_samples 9700 joint_samples 28 [314455, 1041374] [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1042111, 228550] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1042111, 228550] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1046384, 284034] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1046384, 284034] [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure processed_samples 9800 unjoint_samples 9800 joint_samples 28 [587113, 1046850] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [587113, 1046850] processed_samples 9801 unjoint_samples 9800 joint_samples 28 [719368, 1035681] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1037026, 1012001] processed_samples 9801 unjoint_samples 9800 joint_samples 28 [719368, 1035681] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1048080, 357932] processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1037026, 1012001] processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1048080, 357932] processed_samples 9801 unjoint_samples 9800 joint_samples 29 [1047214, 918703] processed_samples 9801 unjoint_samples 9800 joint_samples 29 [1047214, 918703] processed_samples 9802 unjoint_samples 9800 joint_samples 28 [635576, 1041374] processed_samples 9802 unjoint_samples 9800 joint_samples 28 [635576, 1041374] [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure processed_samples 9900 unjoint_samples 9900 joint_samples 28 [1046384, 592146] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [252377, 1030562] [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1042111, 691204] processed_samples 9901 unjoint_samples 9900 joint_samples 30 [222168, 1037389] [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure processed_samples 9900 unjoint_samples 9900 joint_samples 28 [861589, 1046850] processed_samples 9901 unjoint_samples 9900 joint_samples 28 [1010881, 1035681] processed_samples 9900 unjoint_samples 9900 joint_samples 28 [1046384, 592146] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [252377, 1030562] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1048080, 677917] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1042111, 691204] processed_samples 9901 unjoint_samples 9900 joint_samples 30 [222168, 1037389] processed_samples 9900 unjoint_samples 9900 joint_samples 28 [861589, 1046850] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 9902 unjoint_samples 9900 joint_samples 28 [1039122, 1041374] processed_samples 9901 unjoint_samples 9900 joint_samples 28 [1010881, 1035681] processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1048080, 677917] [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure processed_samples 9902 unjoint_samples 9900 joint_samples 28 [1039122, 1041374] [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1048080, 1007239] [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1048080, 1007239] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1044773, 134080] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1044773, 134080] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [559071, 1030562] processed_samples 10001 unjoint_samples 10000 joint_samples 29 [270012, 1045016] processed_samples 10001 unjoint_samples 10000 joint_samples 30 [513749, 1037389] processed_samples 10001 unjoint_samples 10000 joint_samples 29 [270012, 1045016] processed_samples 10000 unjoint_samples 10000 joint_samples 28 [1046384, 986866] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1042111, 953016] processed_samples 10001 unjoint_samples 10000 joint_samples 30 [513749, 1037389] processed_samples 10000 unjoint_samples 10000 joint_samples 29 [559071, 1030562] [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1042111, 953016] processed_samples 10000 unjoint_samples 10000 joint_samples 28 [1046384, 986866] [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure processed_samples 10002 unjoint_samples 10000 joint_samples 29 [1048013, 383798] processed_samples 10002 unjoint_samples 10000 joint_samples 29 [1048013, 383798] [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure processed_samples 10100 unjoint_samples 10100 joint_samples 29 [1044773, 423870] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [1044773, 423870] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [1046316, 166463] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [1046316, 166463] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [297968, 1040092] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [297968, 1040092] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [884096, 1030562] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [246873, 1046423] processed_samples 10100 unjoint_samples 10100 joint_samples 30 [246873, 1046423] processed_samples 10100 unjoint_samples 10100 joint_samples 29 [884096, 1030562] processed_samples 10101 unjoint_samples 10100 joint_samples 29 [538712, 1045016] processed_samples 10101 unjoint_samples 10100 joint_samples 29 [538712, 1045016] processed_samples 10101 unjoint_samples 10100 joint_samples 30 [760695, 1037389] processed_samples 10101 unjoint_samples 10100 joint_samples 30 [760695, 1037389] processed_samples 10102 unjoint_samples 10100 joint_samples 29 [1048013, 722395] processed_samples 10102 unjoint_samples 10100 joint_samples 29 [1048013, 722395] [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1046316, 461146] processed_samples 10200 unjoint_samples 10200 joint_samples 29 [552591, 1040092] [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure processed_samples 10200 unjoint_samples 10200 joint_samples 29 [1044773, 682706] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [591758, 1046423] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [93584, 1045183] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1046316, 461146] processed_samples 10200 unjoint_samples 10200 joint_samples 29 [552591, 1040092] [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure processed_samples 10201 unjoint_samples 10200 joint_samples 29 [854553, 1045016] processed_samples 10200 unjoint_samples 10200 joint_samples 29 [1044773, 682706] processed_samples 10201 unjoint_samples 10200 joint_samples 30 [1001872, 1037389] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [93584, 1045183] processed_samples 10200 unjoint_samples 10200 joint_samples 30 [591758, 1046423] processed_samples 10201 unjoint_samples 10200 joint_samples 29 [854553, 1045016] processed_samples 10201 unjoint_samples 10200 joint_samples 30 [1001872, 1037389] processed_samples 10202 unjoint_samples 10200 joint_samples 30 [1048013, 9045] processed_samples 10202 unjoint_samples 10200 joint_samples 30 [1048013, 9045] [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6ce619240] mmco: unref short failure [h264 @ 0x55e6ce619240] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure processed_samples 10300 unjoint_samples 10300 joint_samples 30 [421159, 1045183] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [421159, 1045183] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1046316, 711195] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1046316, 711195] processed_samples 10300 unjoint_samples 10300 joint_samples 29 [936904, 1040092] processed_samples 10300 unjoint_samples 10300 joint_samples 29 [936904, 1040092] processed_samples 10300 unjoint_samples 10300 joint_samples 29 [1044773, 981784] processed_samples 10300 unjoint_samples 10300 joint_samples 29 [1044773, 981784] processed_samples 10301 unjoint_samples 10300 joint_samples 31 [1033730, 249971] processed_samples 10301 unjoint_samples 10300 joint_samples 31 [1033730, 249971] processed_samples 10301 unjoint_samples 10300 joint_samples 30 [1047040, 91990] processed_samples 10301 unjoint_samples 10300 joint_samples 30 [1047040, 91990] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1009257, 1046423] processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1009257, 1046423] processed_samples 10302 unjoint_samples 10300 joint_samples 30 [1048013, 301246] processed_samples 10302 unjoint_samples 10300 joint_samples 30 [1048013, 301246] [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4ca4640] [h264 @ 0x55e6d1725dc0] mmco: unref short failure mmco: unref short failure [h264 @ 0x55e6d1725dc0] [h264 @ 0x561ea4ca4640] mmco: unref short failure mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1046676, 180019] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1046676, 180019] processed_samples 10400 unjoint_samples 10400 joint_samples 31 [1046316, 109755] processed_samples 10400 unjoint_samples 10400 joint_samples 31 [1046316, 109755] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [297416, 998330] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [297416, 998330] processed_samples 10400 unjoint_samples 10400 joint_samples 31 [288701, 1046423] [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure processed_samples 10400 unjoint_samples 10400 joint_samples 31 [288701, 1046423] processed_samples 10401 unjoint_samples 10400 joint_samples 30 [1047040, 354334] processed_samples 10401 unjoint_samples 10400 joint_samples 30 [1047040, 354334] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [720538, 1045183] processed_samples 10402 unjoint_samples 10400 joint_samples 30 [1048013, 641493] processed_samples 10400 unjoint_samples 10400 joint_samples 30 [720538, 1045183] processed_samples 10402 unjoint_samples 10400 joint_samples 30 [1048013, 641493] processed_samples 10401 unjoint_samples 10400 joint_samples 31 [1033730, 573888] processed_samples 10401 unjoint_samples 10400 joint_samples 31 [1033730, 573888] [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure processed_samples 10500 unjoint_samples 10500 joint_samples 30 [1046676, 493189] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [1046316, 400156] processed_samples 10500 unjoint_samples 10500 joint_samples 30 [564957, 998330] processed_samples 10501 unjoint_samples 10500 joint_samples 30 [1047040, 627299] processed_samples 10500 unjoint_samples 10500 joint_samples 30 [1000083, 1045183] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [612481, 1046423] processed_samples 10501 unjoint_samples 10500 joint_samples 31 [1033730, 824067] processed_samples 10500 unjoint_samples 10500 joint_samples 30 [1046676, 493189] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [1046316, 400156] processed_samples 10500 unjoint_samples 10500 joint_samples 30 [564957, 998330] processed_samples 10501 unjoint_samples 10500 joint_samples 30 [1047040, 627299] processed_samples 10500 unjoint_samples 10500 joint_samples 31 [612481, 1046423] processed_samples 10501 unjoint_samples 10500 joint_samples 31 [1033730, 824067] processed_samples 10500 unjoint_samples 10500 joint_samples 30 [1000083, 1045183] processed_samples 10502 unjoint_samples 10500 joint_samples 30 [1048013, 979835] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure processed_samples 10502 unjoint_samples 10500 joint_samples 30 [1048013, 979835] [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure processed_samples 10600 unjoint_samples 10600 joint_samples 31 [296825, 1045486] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [296825, 1045486] processed_samples 10600 unjoint_samples 10600 joint_samples 30 [1046676, 718683] processed_samples 10600 unjoint_samples 10600 joint_samples 30 [1046676, 718683] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1046316, 731822] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [1046316, 731822] [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure processed_samples 10600 unjoint_samples 10600 joint_samples 31 [876920, 1046423] processed_samples 10600 unjoint_samples 10600 joint_samples 31 [876920, 1046423] processed_samples 10601 unjoint_samples 10600 joint_samples 32 [1037290, 58814] processed_samples 10601 unjoint_samples 10600 joint_samples 32 [1037290, 58814] processed_samples 10600 unjoint_samples 10600 joint_samples 30 [789645, 998330] processed_samples 10602 unjoint_samples 10600 joint_samples 31 [354671, 1019626] processed_samples 10600 unjoint_samples 10600 joint_samples 30 [789645, 998330] processed_samples 10602 unjoint_samples 10600 joint_samples 31 [354671, 1019626] processed_samples 10601 unjoint_samples 10600 joint_samples 30 [1047040, 911473] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure processed_samples 10601 unjoint_samples 10600 joint_samples 30 [1047040, 911473] [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 10700 unjoint_samples 10700 joint_samples 32 [83997, 1031591] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [130088, 960814] processed_samples 10700 unjoint_samples 10700 joint_samples 32 [1046518, 113704] processed_samples 10700 unjoint_samples 10700 joint_samples 32 [83997, 1031591] processed_samples 10700 unjoint_samples 10700 joint_samples 32 [1046518, 113704] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [130088, 960814] processed_samples 10701 unjoint_samples 10700 joint_samples 31 [181777, 1043006] processed_samples 10701 unjoint_samples 10700 joint_samples 31 [181777, 1043006] processed_samples 10701 unjoint_samples 10700 joint_samples 32 [1037290, 332184] processed_samples 10701 unjoint_samples 10700 joint_samples 32 [1037290, 332184] processed_samples 10700 unjoint_samples 10700 joint_samples 30 [1007542, 1008581] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [539938, 1045486] processed_samples 10700 unjoint_samples 10700 joint_samples 31 [539938, 1045486] processed_samples 10700 unjoint_samples 10700 joint_samples 30 [1007542, 1008581] processed_samples 10702 unjoint_samples 10700 joint_samples 31 [650341, 1019626] processed_samples 10702 unjoint_samples 10700 joint_samples 31 [650341, 1019626] [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea23d1800] [h264 @ 0x55e6ce23a1c0] mmco: unref short failure mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure processed_samples 10800 unjoint_samples 10800 joint_samples 31 [1036988, 249709] processed_samples 10800 unjoint_samples 10800 joint_samples 31 [1036988, 249709] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [410360, 1031591] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [410360, 1031591] processed_samples 10800 unjoint_samples 10800 joint_samples 31 [402268, 960814] processed_samples 10800 unjoint_samples 10800 joint_samples 31 [402268, 960814] [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure processed_samples 10800 unjoint_samples 10800 joint_samples 31 [952637, 1045486] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [1046518, 350174] processed_samples 10800 unjoint_samples 10800 joint_samples 32 [1046518, 350174] processed_samples 10801 unjoint_samples 10800 joint_samples 31 [499209, 1043006] processed_samples 10800 unjoint_samples 10800 joint_samples 31 [952637, 1045486] processed_samples 10801 unjoint_samples 10800 joint_samples 31 [499209, 1043006] processed_samples 10801 unjoint_samples 10800 joint_samples 32 [1037290, 611392] processed_samples 10801 unjoint_samples 10800 joint_samples 32 [1037290, 611392] processed_samples 10802 unjoint_samples 10800 joint_samples 31 [910532, 1019626] processed_samples 10802 unjoint_samples 10800 joint_samples 31 [910532, 1019626] [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea1f53700] mmco: unref short failure [h264 @ 0x561ea1f53700] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure processed_samples 10900 unjoint_samples 10900 joint_samples 31 [1036988, 575519] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [769072, 1031591]processed_samples 10900 unjoint_samples 10900 joint_samples 32 [769072, 1031591] processed_samples 10900 unjoint_samples 10900 joint_samples 31 [1036988, 575519] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [207433, 1047330] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [1046518, 649214] processed_samples 10900 unjoint_samples 10900 joint_samples 31 [727333, 960814] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [207433, 1047330] processed_samples 10900 unjoint_samples 10900 joint_samples 32 [1046518, 649214] processed_samples 10902 unjoint_samples 10900 joint_samples 32 [1047105, 32861] processed_samples 10900 unjoint_samples 10900 joint_samples 31 [727333, 960814] processed_samples 10901 unjoint_samples 10900 joint_samples 32 [1037290, 949388] processed_samples 10902 unjoint_samples 10900 joint_samples 32 [1047105, 32861] processed_samples 10901 unjoint_samples 10900 joint_samples 31 [773480, 1043006] processed_samples 10901 unjoint_samples 10900 joint_samples 32 [1037290, 949388] [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure processed_samples 10901 unjoint_samples 10900 joint_samples 31 [773480, 1043006] [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea94d7900] stream 1, offset 0x1400056: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6ce06a440] stream 1, offset 0x1400056: partial file [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure processed_samples 11000 unjoint_samples 11000 joint_samples 32 [508329, 1047330] processed_samples 11000 unjoint_samples 11000 joint_samples 31 [1036988, 843479] processed_samples 11000 unjoint_samples 11000 joint_samples 31 [1016097, 1014030] processed_samples 11001 unjoint_samples 11000 joint_samples 33 [1046543, 164324] processed_samples 11000 unjoint_samples 11000 joint_samples 32 [1006606, 1031591] processed_samples 11002 unjoint_samples 11000 joint_samples 32 [1047105, 325275] processed_samples 11001 unjoint_samples 11000 joint_samples 31 [1020102, 1043006] processed_samples 11001 unjoint_samples 11000 joint_samples 32 [1046518, 887077] [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure processed_samples 11000 unjoint_samples 11000 joint_samples 32 [508329, 1047330] processed_samples 11000 unjoint_samples 11000 joint_samples 31 [1036988, 843479] processed_samples 11001 unjoint_samples 11000 joint_samples 33 [1046543, 164324] processed_samples 11000 unjoint_samples 11000 joint_samples 31 [1016097, 1014030] processed_samples 11000 unjoint_samples 11000 joint_samples 32 [1006606, 1031591] processed_samples 11002 unjoint_samples 11000 joint_samples 32 [1047105, 325275] processed_samples 11001 unjoint_samples 11000 joint_samples 32 [1046518, 887077] processed_samples 11001 unjoint_samples 11000 joint_samples 31 [1020102, 1043006] [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure processed_samples 11100 unjoint_samples 11100 joint_samples 33 [169625, 1045617] processed_samples 11100 unjoint_samples 11100 joint_samples 33 [169625, 1045617] processed_samples 11100 unjoint_samples 11100 joint_samples 32 [98042, 1030253] processed_samples 11100 unjoint_samples 11100 joint_samples 32 [98042, 1030253] processed_samples 11100 unjoint_samples 11100 joint_samples 32 [511764, 1016325] processed_samples 11100 unjoint_samples 11100 joint_samples 32 [511764, 1016325] [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure processed_samples 11101 unjoint_samples 11100 joint_samples 32 [1042397, 327835] processed_samples 11101 unjoint_samples 11100 joint_samples 32 [1042397, 327835] processed_samples 11101 unjoint_samples 11100 joint_samples 33 [1046543, 462595] [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure processed_samples 11101 unjoint_samples 11100 joint_samples 33 [1046543, 462595] [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure processed_samples 11101 unjoint_samples 11100 joint_samples 33 [118142, 1044086] processed_samples 11101 unjoint_samples 11100 joint_samples 33 [118142, 1044086] processed_samples 11100 unjoint_samples 11100 joint_samples 32 [842925, 1047330] processed_samples 11100 unjoint_samples 11100 joint_samples 32 [842925, 1047330] processed_samples 11102 unjoint_samples 11100 joint_samples 32 [1047105, 548011] processed_samples 11102 unjoint_samples 11100 joint_samples 32 [1047105, 548011] [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure processed_samples 11200 unjoint_samples 11200 joint_samples 33 [423324, 1045617] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1040249, 152913] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [423324, 1045617] processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1040249, 152913] processed_samples 11200 unjoint_samples 11200 joint_samples 32 [451236, 1030253] processed_samples 11200 unjoint_samples 11200 joint_samples 32 [451236, 1030253] processed_samples 11200 unjoint_samples 11200 joint_samples 32 [769052, 1016325] processed_samples 11200 unjoint_samples 11200 joint_samples 32 [769052, 1016325] processed_samples 11201 unjoint_samples 11200 joint_samples 33 [1046543, 855053] processed_samples 11201 unjoint_samples 11200 joint_samples 32 [1042397, 606253] processed_samples 11201 unjoint_samples 11200 joint_samples 33 [1046543, 855053] processed_samples 11201 unjoint_samples 11200 joint_samples 32 [1042397, 606253] processed_samples 11201 unjoint_samples 11200 joint_samples 33 [390273, 1044086] processed_samples 11201 unjoint_samples 11200 joint_samples 33 [390273, 1044086] processed_samples 11202 unjoint_samples 11200 joint_samples 32 [1047105, 870635] processed_samples 11202 unjoint_samples 11200 joint_samples 32 [1047105, 870635] [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 11300 unjoint_samples 11300 joint_samples 33 [746581, 1045617] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [746581, 1045617] processed_samples 11301 unjoint_samples 11300 joint_samples 33 [69333, 1036568] [h264 @ 0x55e6cea6c680] mmco: unref short failure processed_samples 11301 unjoint_samples 11300 joint_samples 33 [69333, 1036568] [h264 @ 0x561ea4c60e40] mmco: unref short failure processed_samples 11300 unjoint_samples 11300 joint_samples 32 [774809, 1030253] processed_samples 11300 unjoint_samples 11300 joint_samples 32 [774809, 1030253] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1040249, 492733] processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1040249, 492733] processed_samples 11300 unjoint_samples 11300 joint_samples 32 [1017287, 1018752] processed_samples 11301 unjoint_samples 11300 joint_samples 34 [207444, 1039783] processed_samples 11300 unjoint_samples 11300 joint_samples 32 [1017287, 1018752] [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure processed_samples 11301 unjoint_samples 11300 joint_samples 34 [207444, 1039783] processed_samples 11302 unjoint_samples 11300 joint_samples 33 [1048067, 190609] [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure processed_samples 11301 unjoint_samples 11300 joint_samples 33 [710115, 1044086] processed_samples 11302 unjoint_samples 11300 joint_samples 33 [1048067, 190609] processed_samples 11301 unjoint_samples 11300 joint_samples 33 [710115, 1044086] [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1038212, 299902] processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1038212, 299902] processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1040249, 799327] processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1007738, 1045617] processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1040249, 799327] processed_samples 11401 unjoint_samples 11400 joint_samples 34 [524738, 1039783] processed_samples 11400 unjoint_samples 11400 joint_samples 32 [1022305, 1030253] processed_samples 11401 unjoint_samples 11400 joint_samples 34 [524738, 1039783] processed_samples 11400 unjoint_samples 11400 joint_samples 32 [1022305, 1030253] processed_samples 11401 unjoint_samples 11400 joint_samples 33 [370444, 1036568] processed_samples 11401 unjoint_samples 11400 joint_samples 33 [370444, 1036568] processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1007738, 1045617] processed_samples 11402 unjoint_samples 11400 joint_samples 33 [1048067, 479527] processed_samples 11402 unjoint_samples 11400 joint_samples 33 [1048067, 479527] processed_samples 11401 unjoint_samples 11400 joint_samples 33 [1028292, 1044086] processed_samples 11401 unjoint_samples 11400 joint_samples 33 [1028292, 1044086] [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea4dcf140] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea053a580] stream 1, offset 0xbcbd: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea053a580] stream 1, offset 0xbe30: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6d5d311c0] stream 1, offset 0xbcbd: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6d5d311c0] stream 1, offset 0xbe30: partial file [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure processed_samples 11500 unjoint_samples 11500 joint_samples 33 [293873, 1040917] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [1012232, 365358] processed_samples 11500 unjoint_samples 11500 joint_samples 33 [293873, 1040917] processed_samples 11500 unjoint_samples 11500 joint_samples 34 [1012232, 365358] [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure processed_samples 11500 unjoint_samples 11500 joint_samples 33 [1038212, 689517] processed_samples 11500 unjoint_samples 11500 joint_samples 33 [1038212, 689517] processed_samples 11501 unjoint_samples 11500 joint_samples 34 [24900, 1046655] processed_samples 11501 unjoint_samples 11500 joint_samples 34 [24900, 1046655] processed_samples 11501 unjoint_samples 11500 joint_samples 34 [1034782, 351192] processed_samples 11501 unjoint_samples 11500 joint_samples 34 [1034782, 351192] processed_samples 11501 unjoint_samples 11500 joint_samples 33 [691720, 1036568] processed_samples 11501 unjoint_samples 11500 joint_samples 33 [691720, 1036568] processed_samples 11502 unjoint_samples 11500 joint_samples 33 [1048067, 765432] processed_samples 11502 unjoint_samples 11500 joint_samples 33 [1048067, 765432] [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure processed_samples 11501 unjoint_samples 11500 joint_samples 34 [829223, 1039783] processed_samples 11501 unjoint_samples 11500 joint_samples 34 [829223, 1039783] [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea45d3300] [h264 @ 0x55e6d6193600] mmco: unref short failure mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure processed_samples 11600 unjoint_samples 11600 joint_samples 34 [1012232, 660379] processed_samples 11600 unjoint_samples 11600 joint_samples 34 [1012232, 660379] processed_samples 11600 unjoint_samples 11600 joint_samples 33 [621171, 1040917] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [1043448, 1043131] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [1034782, 793343] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [289901, 1046655] processed_samples 11600 unjoint_samples 11600 joint_samples 33 [621171, 1040917] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [1034782, 793343] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [1041360, 16150] processed_samples 11600 unjoint_samples 11600 joint_samples 33 [1038212, 980949] processed_samples 11602 unjoint_samples 11600 joint_samples 34 [169912, 1027423] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [289901, 1046655] processed_samples 11600 unjoint_samples 11600 joint_samples 33 [1038212, 980949] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [1043448, 1043131] processed_samples 11602 unjoint_samples 11600 joint_samples 34 [169912, 1027423] processed_samples 11601 unjoint_samples 11600 joint_samples 34 [1041360, 16150] [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure processed_samples 11700 unjoint_samples 11700 joint_samples 34 [1047469, 233881] processed_samples 11700 unjoint_samples 11700 joint_samples 34 [1047469, 233881] processed_samples 11700 unjoint_samples 11700 joint_samples 33 [1042357, 1040917] processed_samples 11700 unjoint_samples 11700 joint_samples 33 [1042357, 1040917] processed_samples 11701 unjoint_samples 11700 joint_samples 35 [1047154, 36648] processed_samples 11701 unjoint_samples 11700 joint_samples 35 [1047154, 36648] processed_samples 11701 unjoint_samples 11700 joint_samples 35 [302300, 1045414] processed_samples 11701 unjoint_samples 11700 joint_samples 35 [302300, 1045414] processed_samples 11701 unjoint_samples 11700 joint_samples 34 [1041360, 325251] processed_samples 11701 unjoint_samples 11700 joint_samples 34 [1041360, 325251] processed_samples 11701 unjoint_samples 11700 joint_samples 34 [631408, 1046655] processed_samples 11701 unjoint_samples 11700 joint_samples 34 [631408, 1046655] processed_samples 11700 unjoint_samples 11700 joint_samples 34 [1025746, 1036665] processed_samples 11700 unjoint_samples 11700 joint_samples 34 [1025746, 1036665] processed_samples 11702 unjoint_samples 11700 joint_samples 34 [491622, 1027423] processed_samples 11702 unjoint_samples 11700 joint_samples 34 [491622, 1027423] [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 11800 unjoint_samples 11800 joint_samples 34 [204126, 1046979] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [1038775, 373423] processed_samples 11800 unjoint_samples 11800 joint_samples 34 [1047469, 533790] processed_samples 11801 unjoint_samples 11800 joint_samples 35 [744382, 1045414] processed_samples 11800 unjoint_samples 11800 joint_samples 34 [204126, 1046979] processed_samples 11800 unjoint_samples 11800 joint_samples 35 [1038775, 373423] processed_samples 11801 unjoint_samples 11800 joint_samples 35 [1047154, 444342] processed_samples 11801 unjoint_samples 11800 joint_samples 35 [744382, 1045414] processed_samples 11800 unjoint_samples 11800 joint_samples 34 [1047469, 533790] processed_samples 11801 unjoint_samples 11800 joint_samples 35 [1047154, 444342] processed_samples 11801 unjoint_samples 11800 joint_samples 34 [931296, 1046655] processed_samples 11801 unjoint_samples 11800 joint_samples 34 [1041360, 610025] [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure processed_samples 11801 unjoint_samples 11800 joint_samples 34 [931296, 1046655] [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure processed_samples 11801 unjoint_samples 11800 joint_samples 34 [1041360, 610025] processed_samples 11802 unjoint_samples 11800 joint_samples 34 [955524, 1027423] [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure processed_samples 11802 unjoint_samples 11800 joint_samples 34 [955524, 1027423] [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure processed_samples 11900 unjoint_samples 11900 joint_samples 34 [503374, 1046979] processed_samples 11900 unjoint_samples 11900 joint_samples 34 [503374, 1046979] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1038775, 777589] [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure processed_samples 11901 unjoint_samples 11900 joint_samples 34 [1041360, 850782] processed_samples 11901 unjoint_samples 11900 joint_samples 34 [1041360, 850782] processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1038775, 777589] processed_samples 11901 unjoint_samples 11900 joint_samples 36 [61172, 1045414] processed_samples 11901 unjoint_samples 11900 joint_samples 36 [61172, 1045414] processed_samples 11901 unjoint_samples 11900 joint_samples 35 [134588, 1046655] processed_samples 11901 unjoint_samples 11900 joint_samples 35 [134588, 1046655] processed_samples 11901 unjoint_samples 11900 joint_samples 35 [1047154, 677768] processed_samples 11900 unjoint_samples 11900 joint_samples 34 [1047469, 867142] processed_samples 11900 unjoint_samples 11900 joint_samples 34 [1047469, 867142] processed_samples 11902 unjoint_samples 11900 joint_samples 35 [310675, 1027423] processed_samples 11902 unjoint_samples 11900 joint_samples 35 [310675, 1027423] processed_samples 11901 unjoint_samples 11900 joint_samples 35 [1047154, 677768] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 12000 unjoint_samples 12000 joint_samples 36 [1046385, 52681] processed_samples 12000 unjoint_samples 12000 joint_samples 36 [1046385, 52681] processed_samples 12000 unjoint_samples 12000 joint_samples 35 [206835, 1043335] processed_samples 12000 unjoint_samples 12000 joint_samples 35 [206835, 1043335] processed_samples 12001 unjoint_samples 12000 joint_samples 35 [176926, 1045289] processed_samples 12001 unjoint_samples 12000 joint_samples 35 [176926, 1045289] [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure processed_samples 12001 unjoint_samples 12000 joint_samples 35 [449449, 1046655] processed_samples 12001 unjoint_samples 12000 joint_samples 35 [449449, 1046655] processed_samples 12000 unjoint_samples 12000 joint_samples 34 [830755, 1046979] processed_samples 12000 unjoint_samples 12000 joint_samples 34 [830755, 1046979] processed_samples 12001 unjoint_samples 12000 joint_samples 35 [1047154, 976629] processed_samples 12001 unjoint_samples 12000 joint_samples 35 [1047154, 976629] [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure processed_samples 12001 unjoint_samples 12000 joint_samples 36 [428112, 1045414] processed_samples 12001 unjoint_samples 12000 joint_samples 36 [428112, 1045414] processed_samples 12002 unjoint_samples 12000 joint_samples 35 [574999, 1027423] processed_samples 12002 unjoint_samples 12000 joint_samples 35 [574999, 1027423] [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure processed_samples 12100 unjoint_samples 12100 joint_samples 35 [1045413, 174955] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [1046385, 294968] processed_samples 12100 unjoint_samples 12100 joint_samples 35 [1045413, 174955] processed_samples 12100 unjoint_samples 12100 joint_samples 36 [1046385, 294968] processed_samples 12100 unjoint_samples 12100 joint_samples 35 [599074, 1043335] processed_samples 12100 unjoint_samples 12100 joint_samples 35 [599074, 1043335] processed_samples 12101 unjoint_samples 12100 joint_samples 35 [652261, 1045289] processed_samples 12101 unjoint_samples 12100 joint_samples 35 [652261, 1045289] processed_samples 12101 unjoint_samples 12100 joint_samples 35 [764341, 1046655] processed_samples 12101 unjoint_samples 12100 joint_samples 36 [731288, 1045414] processed_samples 12101 unjoint_samples 12100 joint_samples 36 [402257, 1031111] processed_samples 12101 unjoint_samples 12100 joint_samples 36 [731288, 1045414] processed_samples 12101 unjoint_samples 12100 joint_samples 35 [764341, 1046655] processed_samples 12101 unjoint_samples 12100 joint_samples 36 [402257, 1031111] processed_samples 12102 unjoint_samples 12100 joint_samples 35 [884095, 1027423] [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure processed_samples 12102 unjoint_samples 12100 joint_samples 35 [884095, 1027423] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cec12f80] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure processed_samples 12200 unjoint_samples 12200 joint_samples 35 [1045413, 474640] [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure processed_samples 12200 unjoint_samples 12200 joint_samples 35 [1045413, 474640] [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure processed_samples 12200 unjoint_samples 12200 joint_samples 36 [1046385, 653747] processed_samples 12200 unjoint_samples 12200 joint_samples 36 [1046385, 653747] processed_samples 12201 unjoint_samples 12200 joint_samples 35 [937018, 1045289] processed_samples 12201 unjoint_samples 12200 joint_samples 36 [105740, 1046655] processed_samples 12200 unjoint_samples 12200 joint_samples 35 [876262, 1043335] processed_samples 12201 unjoint_samples 12200 joint_samples 35 [937018, 1045289] processed_samples 12201 unjoint_samples 12200 joint_samples 36 [105740, 1046655] processed_samples 12200 unjoint_samples 12200 joint_samples 35 [876262, 1043335] processed_samples 12201 unjoint_samples 12200 joint_samples 36 [736217, 1031111] processed_samples 12202 unjoint_samples 12200 joint_samples 36 [1047555, 98416] processed_samples 12201 unjoint_samples 12200 joint_samples 36 [736217, 1031111] processed_samples 12201 unjoint_samples 12200 joint_samples 36 [1010831, 1045414] processed_samples 12202 unjoint_samples 12200 joint_samples 36 [1047555, 98416] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure processed_samples 12201 unjoint_samples 12200 joint_samples 36 [1010831, 1045414] [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea75ecf00] mmco: unref short failure [h264 @ 0x561ea75ecf00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure processed_samples 12300 unjoint_samples 12300 joint_samples 36 [1047354, 108614] processed_samples 12300 unjoint_samples 12300 joint_samples 36 [1047354, 108614] processed_samples 12300 unjoint_samples 12300 joint_samples 35 [1045413, 878898] processed_samples 12300 unjoint_samples 12300 joint_samples 35 [1045413, 878898] processed_samples 12301 unjoint_samples 12300 joint_samples 36 [1043817, 163441] [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure processed_samples 12301 unjoint_samples 12300 joint_samples 36 [1043817, 163441] [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure processed_samples 12301 unjoint_samples 12300 joint_samples 37 [142980, 1046618] processed_samples 12301 unjoint_samples 12300 joint_samples 37 [142980, 1046618] processed_samples 12301 unjoint_samples 12300 joint_samples 36 [417626, 1046655] processed_samples 12301 unjoint_samples 12300 joint_samples 36 [417626, 1046655] processed_samples 12300 unjoint_samples 12300 joint_samples 36 [1046385, 904768] processed_samples 12300 unjoint_samples 12300 joint_samples 36 [1046385, 904768] processed_samples 12301 unjoint_samples 12300 joint_samples 37 [1039781, 343251] processed_samples 12301 unjoint_samples 12300 joint_samples 37 [1039781, 343251] processed_samples 12302 unjoint_samples 12300 joint_samples 36 [1047555, 359698] processed_samples 12302 unjoint_samples 12300 joint_samples 36 [1047555, 359698] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure processed_samples 12400 unjoint_samples 12400 joint_samples 37 [129718, 1041510] processed_samples 12400 unjoint_samples 12400 joint_samples 37 [129718, 1041510] processed_samples 12400 unjoint_samples 12400 joint_samples 36 [155990, 1023638] processed_samples 12400 unjoint_samples 12400 joint_samples 36 [155990, 1023638] processed_samples 12400 unjoint_samples 12400 joint_samples 36 [1047354, 398333] processed_samples 12400 unjoint_samples 12400 joint_samples 36 [1047354, 398333] processed_samples 12401 unjoint_samples 12400 joint_samples 37 [492416, 1046618] processed_samples 12401 unjoint_samples 12400 joint_samples 37 [492416, 1046618] processed_samples 12401 unjoint_samples 12400 joint_samples 36 [1043817, 481497] processed_samples 12401 unjoint_samples 12400 joint_samples 36 [1043817, 481497] processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1039781, 802263] processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1039781, 802263] processed_samples 12402 unjoint_samples 12400 joint_samples 36 [1047555, 672540] processed_samples 12402 unjoint_samples 12400 joint_samples 36 [1047555, 672540] processed_samples 12401 unjoint_samples 12400 joint_samples 36 [836216, 1046655] processed_samples 12401 unjoint_samples 12400 joint_samples 36 [836216, 1046655] [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea0919140] stream 1, offset 0x1400a4d: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6cf3a7980] stream 1, offset 0x1400a4d: partial file [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure processed_samples 12500 unjoint_samples 12500 joint_samples 36 [512120, 1023638] processed_samples 12500 unjoint_samples 12500 joint_samples 36 [512120, 1023638] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [426622, 1041510] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [426622, 1041510] processed_samples 12501 unjoint_samples 12500 joint_samples 38 [1046420, 22519] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1029669, 101040] processed_samples 12501 unjoint_samples 12500 joint_samples 38 [1046420, 22519] processed_samples 12500 unjoint_samples 12500 joint_samples 36 [1047354, 717623] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1029669, 101040] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [807400, 1046618] processed_samples 12500 unjoint_samples 12500 joint_samples 36 [1047354, 717623] processed_samples 12502 unjoint_samples 12500 joint_samples 37 [243099, 973385] processed_samples 12501 unjoint_samples 12500 joint_samples 37 [807400, 1046618] processed_samples 12501 unjoint_samples 12500 joint_samples 36 [1043817, 741041] processed_samples 12502 unjoint_samples 12500 joint_samples 37 [243099, 973385] processed_samples 12501 unjoint_samples 12500 joint_samples 36 [1043817, 741041] [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure processed_samples 12600 unjoint_samples 12600 joint_samples 37 [119783, 1041803] processed_samples 12600 unjoint_samples 12600 joint_samples 37 [119783, 1041803] processed_samples 12600 unjoint_samples 12600 joint_samples 36 [947528, 1023638] processed_samples 12601 unjoint_samples 12600 joint_samples 37 [688323, 1041510] processed_samples 12601 unjoint_samples 12600 joint_samples 37 [688323, 1041510] processed_samples 12600 unjoint_samples 12600 joint_samples 36 [947528, 1023638] processed_samples 12601 unjoint_samples 12600 joint_samples 37 [1029669, 508464] processed_samples 12601 unjoint_samples 12600 joint_samples 37 [1029669, 508464] processed_samples 12601 unjoint_samples 12600 joint_samples 38 [1046420, 329783] processed_samples 12601 unjoint_samples 12600 joint_samples 38 [1046420, 329783] processed_samples 12601 unjoint_samples 12600 joint_samples 38 [1043956, 90917] processed_samples 12601 unjoint_samples 12600 joint_samples 38 [1043956, 90917] [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure processed_samples 12602 unjoint_samples 12600 joint_samples 37 [552149, 973385] processed_samples 12602 unjoint_samples 12600 joint_samples 37 [552149, 973385] processed_samples 12601 unjoint_samples 12600 joint_samples 36 [1043817, 1035203] processed_samples 12601 unjoint_samples 12600 joint_samples 36 [1043817, 1035203] [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cf2ed600] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure processed_samples 12700 unjoint_samples 12700 joint_samples 37 [208263, 1025290] processed_samples 12700 unjoint_samples 12700 joint_samples 37 [208263, 1025290] processed_samples 12700 unjoint_samples 12700 joint_samples 37 [385554, 1041803] processed_samples 12701 unjoint_samples 12700 joint_samples 38 [1046420, 665517] [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure processed_samples 12701 unjoint_samples 12700 joint_samples 37 [249902, 1047194] processed_samples 12701 unjoint_samples 12700 joint_samples 37 [1029669, 957885] processed_samples 12701 unjoint_samples 12700 joint_samples 37 [1001592, 1041510] processed_samples 12701 unjoint_samples 12700 joint_samples 38 [1043956, 345010] processed_samples 12700 unjoint_samples 12700 joint_samples 37 [385554, 1041803] [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure processed_samples 12701 unjoint_samples 12700 joint_samples 38 [1046420, 665517] processed_samples 12702 unjoint_samples 12700 joint_samples 37 [844195, 973385] [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure processed_samples 12701 unjoint_samples 12700 joint_samples 37 [249902, 1047194] processed_samples 12701 unjoint_samples 12700 joint_samples 37 [1029669, 957885] processed_samples 12701 unjoint_samples 12700 joint_samples 38 [1043956, 345010] processed_samples 12701 unjoint_samples 12700 joint_samples 37 [1001592, 1041510] [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure processed_samples 12702 unjoint_samples 12700 joint_samples 37 [844195, 973385] [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure processed_samples 12800 unjoint_samples 12800 joint_samples 37 [663325, 1041803] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [160825, 1040860] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [160825, 1040860] processed_samples 12800 unjoint_samples 12800 joint_samples 37 [663325, 1041803] processed_samples 12801 unjoint_samples 12800 joint_samples 37 [544946, 1047194] processed_samples 12801 unjoint_samples 12800 joint_samples 37 [544946, 1047194] processed_samples 12800 unjoint_samples 12800 joint_samples 37 [665225, 1025290] processed_samples 12800 unjoint_samples 12800 joint_samples 37 [665225, 1025290] processed_samples 12802 unjoint_samples 12800 joint_samples 38 [13220, 1047130] processed_samples 12802 unjoint_samples 12800 joint_samples 38 [13220, 1047130] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1036633, 323147] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1036633, 323147] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1046420, 1004134] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1046420, 1004134] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1043956, 664689] processed_samples 12801 unjoint_samples 12800 joint_samples 38 [1043956, 664689] [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 12900 unjoint_samples 12900 joint_samples 37 [941589, 1041803] processed_samples 12901 unjoint_samples 12900 joint_samples 39 [543563, 1028347] processed_samples 12900 unjoint_samples 12900 joint_samples 37 [937194, 1025290] processed_samples 12901 unjoint_samples 12900 joint_samples 38 [457986, 1040860] processed_samples 12901 unjoint_samples 12900 joint_samples 38 [1043956, 1012599] processed_samples 12900 unjoint_samples 12900 joint_samples 37 [941589, 1041803] processed_samples 12901 unjoint_samples 12900 joint_samples 38 [1036633, 608434] processed_samples 12901 unjoint_samples 12900 joint_samples 37 [893014, 1047194] processed_samples 12901 unjoint_samples 12900 joint_samples 39 [543563, 1028347] processed_samples 12902 unjoint_samples 12900 joint_samples 38 [280556, 1047130] processed_samples 12900 unjoint_samples 12900 joint_samples 37 [937194, 1025290] processed_samples 12901 unjoint_samples 12900 joint_samples 38 [1036633, 608434] processed_samples 12901 unjoint_samples 12900 joint_samples 38 [457986, 1040860] processed_samples 12901 unjoint_samples 12900 joint_samples 38 [1043956, 1012599] processed_samples 12901 unjoint_samples 12900 joint_samples 37 [893014, 1047194] processed_samples 12902 unjoint_samples 12900 joint_samples 38 [280556, 1047130] [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure processed_samples 13000 unjoint_samples 13000 joint_samples 38 [986188, 290948] processed_samples 13000 unjoint_samples 13000 joint_samples 38 [986188, 290948] [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure processed_samples 13001 unjoint_samples 13000 joint_samples 38 [91253, 1047194] processed_samples 13000 unjoint_samples 13000 joint_samples 38 [372903, 1044657] processed_samples 13000 unjoint_samples 13000 joint_samples 38 [372903, 1044657] processed_samples 13001 unjoint_samples 13000 joint_samples 38 [91253, 1047194] processed_samples 13001 unjoint_samples 13000 joint_samples 39 [297117, 1046530] processed_samples 13001 unjoint_samples 13000 joint_samples 39 [297117, 1046530] processed_samples 13001 unjoint_samples 13000 joint_samples 39 [908881, 1028347] processed_samples 13001 unjoint_samples 13000 joint_samples 39 [908881, 1028347] processed_samples 13001 unjoint_samples 13000 joint_samples 38 [771142, 1040860] processed_samples 13001 unjoint_samples 13000 joint_samples 38 [771142, 1040860] processed_samples 13002 unjoint_samples 13000 joint_samples 38 [583935, 1047130] processed_samples 13002 unjoint_samples 13000 joint_samples 38 [583935, 1047130] processed_samples 13001 unjoint_samples 13000 joint_samples 38 [1036633, 959323] processed_samples 13001 unjoint_samples 13000 joint_samples 38 [1036633, 959323] [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea5f27180] mmco: unref short failure [h264 @ 0x561ea5f27180] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure processed_samples 13100 unjoint_samples 13100 joint_samples 38 [654265, 1044657] processed_samples 13100 unjoint_samples 13100 joint_samples 38 [654265, 1044657] processed_samples 13101 unjoint_samples 13100 joint_samples 39 [232179, 1039371] processed_samples 13101 unjoint_samples 13100 joint_samples 39 [232179, 1039371] processed_samples 13101 unjoint_samples 13100 joint_samples 38 [335758, 1047194] processed_samples 13101 unjoint_samples 13100 joint_samples 38 [335758, 1047194] processed_samples 13100 unjoint_samples 13100 joint_samples 38 [986188, 524520] processed_samples 13100 unjoint_samples 13100 joint_samples 38 [986188, 524520] processed_samples 13101 unjoint_samples 13100 joint_samples 40 [1030990, 152330] processed_samples 13101 unjoint_samples 13100 joint_samples 40 [1030990, 152330] [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure processed_samples 13101 unjoint_samples 13100 joint_samples 39 [567581, 1046530] processed_samples 13101 unjoint_samples 13100 joint_samples 39 [567581, 1046530] processed_samples 13101 unjoint_samples 13100 joint_samples 38 [1042411, 1043068] processed_samples 13101 unjoint_samples 13100 joint_samples 38 [1042411, 1043068] [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure processed_samples 13102 unjoint_samples 13100 joint_samples 38 [851175, 1047130] processed_samples 13102 unjoint_samples 13100 joint_samples 38 [851175, 1047130] [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure processed_samples 13201 unjoint_samples 13200 joint_samples 39 [515628, 1039371] [h264 @ 0x561ea469e3c0] mmco: unref short failure processed_samples 13201 unjoint_samples 13200 joint_samples 40 [1030990, 487818] processed_samples 13201 unjoint_samples 13200 joint_samples 39 [1042822, 327901] processed_samples 13200 unjoint_samples 13200 joint_samples 38 [948767, 1044657] processed_samples 13201 unjoint_samples 13200 joint_samples 39 [515628, 1039371] processed_samples 13201 unjoint_samples 13200 joint_samples 38 [684165, 1047194] [h264 @ 0x55e6ce98dc00] mmco: unref short failure processed_samples 13201 unjoint_samples 13200 joint_samples 39 [1042822, 327901] processed_samples 13200 unjoint_samples 13200 joint_samples 38 [986188, 831689] processed_samples 13201 unjoint_samples 13200 joint_samples 40 [1030990, 487818] processed_samples 13202 unjoint_samples 13200 joint_samples 39 [81103, 1047427] processed_samples 13200 unjoint_samples 13200 joint_samples 38 [948767, 1044657] processed_samples 13201 unjoint_samples 13200 joint_samples 38 [684165, 1047194] processed_samples 13201 unjoint_samples 13200 joint_samples 39 [865402, 1046530] processed_samples 13200 unjoint_samples 13200 joint_samples 38 [986188, 831689] processed_samples 13201 unjoint_samples 13200 joint_samples 39 [865402, 1046530] processed_samples 13202 unjoint_samples 13200 joint_samples 39 [81103, 1047427] [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1038927, 352466] processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1038927, 352466] processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1035003, 328089] processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1035003, 328089] processed_samples 13301 unjoint_samples 13300 joint_samples 39 [874300, 1039371] processed_samples 13301 unjoint_samples 13300 joint_samples 39 [874300, 1039371] processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1030990, 794119] processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1030990, 794119] processed_samples 13300 unjoint_samples 13300 joint_samples 38 [1040116, 1041003] processed_samples 13300 unjoint_samples 13300 joint_samples 38 [1040116, 1041003] [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 13302 unjoint_samples 13300 joint_samples 39 [352033, 1047427] processed_samples 13301 unjoint_samples 13300 joint_samples 38 [966998, 1047194] processed_samples 13302 unjoint_samples 13300 joint_samples 39 [352033, 1047427] processed_samples 13301 unjoint_samples 13300 joint_samples 38 [966998, 1047194] processed_samples 13301 unjoint_samples 13300 joint_samples 39 [1042822, 598041] processed_samples 13301 unjoint_samples 13300 joint_samples 39 [1042822, 598041] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea45d8e80] mmco: unref short failure [h264 @ 0x561ea45d8e80] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure processed_samples 13401 unjoint_samples 13400 joint_samples 39 [210827, 1047194] processed_samples 13400 unjoint_samples 13400 joint_samples 39 [314407, 1046510] processed_samples 13401 unjoint_samples 13400 joint_samples 39 [210827, 1047194] processed_samples 13400 unjoint_samples 13400 joint_samples 39 [314407, 1046510] processed_samples 13401 unjoint_samples 13400 joint_samples 39 [1042822, 929603] processed_samples 13401 unjoint_samples 13400 joint_samples 39 [1042822, 929603] processed_samples 13401 unjoint_samples 13400 joint_samples 41 [12705, 1046635] processed_samples 13401 unjoint_samples 13400 joint_samples 41 [12705, 1046635] processed_samples 13400 unjoint_samples 13400 joint_samples 39 [1038927, 609705] processed_samples 13400 unjoint_samples 13400 joint_samples 39 [1038927, 609705] processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1046923, 77525] processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1046923, 77525] processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1035003, 839396] processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1035003, 839396] processed_samples 13402 unjoint_samples 13400 joint_samples 39 [661129, 1047427] processed_samples 13402 unjoint_samples 13400 joint_samples 39 [661129, 1047427] [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure processed_samples 13500 unjoint_samples 13500 joint_samples 39 [570198, 1046510] [h264 @ 0x55e6d5b9bf40] mmco: unref short failure processed_samples 13500 unjoint_samples 13500 joint_samples 39 [570198, 1046510] processed_samples 13500 unjoint_samples 13500 joint_samples 39 [1038927, 927561] processed_samples 13500 unjoint_samples 13500 joint_samples 39 [1038927, 927561] processed_samples 13501 unjoint_samples 13500 joint_samples 41 [354543, 1046635] processed_samples 13501 unjoint_samples 13500 joint_samples 40 [1046923, 390458] processed_samples 13501 unjoint_samples 13500 joint_samples 41 [354543, 1046635] processed_samples 13501 unjoint_samples 13500 joint_samples 40 [1046923, 390458] processed_samples 13501 unjoint_samples 13500 joint_samples 39 [558091, 1047194] processed_samples 13501 unjoint_samples 13500 joint_samples 39 [558091, 1047194] processed_samples 13501 unjoint_samples 13500 joint_samples 40 [171588, 1046820] processed_samples 13501 unjoint_samples 13500 joint_samples 40 [171588, 1046820] processed_samples 13501 unjoint_samples 13500 joint_samples 41 [1042101, 93686] processed_samples 13501 unjoint_samples 13500 joint_samples 41 [1042101, 93686] processed_samples 13502 unjoint_samples 13500 joint_samples 39 [958116, 1047427] processed_samples 13502 unjoint_samples 13500 joint_samples 39 [958116, 1047427] [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure processed_samples 13600 unjoint_samples 13600 joint_samples 40 [221341, 1040547] processed_samples 13600 unjoint_samples 13600 joint_samples 40 [221341, 1040547] processed_samples 13601 unjoint_samples 13600 joint_samples 41 [1042101, 373788] processed_samples 13601 unjoint_samples 13600 joint_samples 41 [1042101, 373788] processed_samples 13601 unjoint_samples 13600 joint_samples 40 [454294, 1046820] processed_samples 13601 unjoint_samples 13600 joint_samples 40 [454294, 1046820] processed_samples 13601 unjoint_samples 13600 joint_samples 39 [857958, 1047194] processed_samples 13601 unjoint_samples 13600 joint_samples 39 [857958, 1047194] processed_samples 13600 unjoint_samples 13600 joint_samples 39 [826172, 1046510] processed_samples 13601 unjoint_samples 13600 joint_samples 41 [804088, 1046635] processed_samples 13600 unjoint_samples 13600 joint_samples 39 [826172, 1046510] processed_samples 13601 unjoint_samples 13600 joint_samples 41 [804088, 1046635] processed_samples 13602 unjoint_samples 13600 joint_samples 40 [1017483, 409132] processed_samples 13601 unjoint_samples 13600 joint_samples 40 [1046923, 667004] processed_samples 13601 unjoint_samples 13600 joint_samples 40 [1046923, 667004] processed_samples 13602 unjoint_samples 13600 joint_samples 40 [1017483, 409132] [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea1091680] mmco: unref short failure [h264 @ 0x561ea1091680] mmco: unref short failure [h264 @ 0x55e6d090ef80] mmco: unref short failure [h264 @ 0x55e6d090ef80] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure processed_samples 13700 unjoint_samples 13700 joint_samples 40 [1045576, 77942] [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure processed_samples 13700 unjoint_samples 13700 joint_samples 40 [1045576, 77942] [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure processed_samples 13700 unjoint_samples 13700 joint_samples 40 [588164, 1040547] processed_samples 13700 unjoint_samples 13700 joint_samples 40 [588164, 1040547] processed_samples 13701 unjoint_samples 13700 joint_samples 42 [1035404, 84872] processed_samples 13701 unjoint_samples 13700 joint_samples 42 [1035404, 84872] processed_samples 13701 unjoint_samples 13700 joint_samples 40 [725037, 1046820] processed_samples 13701 unjoint_samples 13700 joint_samples 40 [725037, 1046820] processed_samples 13701 unjoint_samples 13700 joint_samples 40 [1046639, 135569] processed_samples 13701 unjoint_samples 13700 joint_samples 40 [1046639, 135569] processed_samples 13701 unjoint_samples 13700 joint_samples 41 [1042101, 689377] processed_samples 13701 unjoint_samples 13700 joint_samples 41 [1042101, 689377] processed_samples 13701 unjoint_samples 13700 joint_samples 40 [1046923, 950722] processed_samples 13701 unjoint_samples 13700 joint_samples 40 [1046923, 950722] [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure processed_samples 13702 unjoint_samples 13700 joint_samples 40 [1017483, 750140] processed_samples 13702 unjoint_samples 13700 joint_samples 40 [1017483, 750140] [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure processed_samples 13800 unjoint_samples 13800 joint_samples 40 [909941, 1040547] processed_samples 13800 unjoint_samples 13800 joint_samples 40 [909941, 1040547] processed_samples 13801 unjoint_samples 13800 joint_samples 42 [58102, 1043690] processed_samples 13801 unjoint_samples 13800 joint_samples 42 [58102, 1043690] processed_samples 13801 unjoint_samples 13800 joint_samples 40 [1046639, 460020] processed_samples 13801 unjoint_samples 13800 joint_samples 40 [1046639, 460020] processed_samples 13801 unjoint_samples 13800 joint_samples 41 [193596, 1046881] processed_samples 13801 unjoint_samples 13800 joint_samples 41 [193596, 1046881] processed_samples 13800 unjoint_samples 13800 joint_samples 40 [1045576, 343685] processed_samples 13800 unjoint_samples 13800 joint_samples 40 [1045576, 343685] [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 13801 unjoint_samples 13800 joint_samples 42 [1035404, 387232] processed_samples 13801 unjoint_samples 13800 joint_samples 42 [1035404, 387232] processed_samples 13801 unjoint_samples 13800 joint_samples 40 [997211, 1046820] processed_samples 13801 unjoint_samples 13800 joint_samples 40 [997211, 1046820] processed_samples 13802 unjoint_samples 13800 joint_samples 40 [1028604, 1029715] processed_samples 13802 unjoint_samples 13800 joint_samples 40 [1028604, 1029715] [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d5ebea00] mmco: unref short failure [h264 @ 0x55e6d5ebea00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d0971040] [h264 @ 0x561ea3ffc480] mmco: unref short failure mmco: unref short failure [h264 @ 0x55e6d0971040] [h264 @ 0x561ea3ffc480] mmco: unref short failure mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure processed_samples 13900 unjoint_samples 13900 joint_samples 41 [1040737, 379854] processed_samples 13900 unjoint_samples 13900 joint_samples 40 [1045576, 630565] processed_samples 13900 unjoint_samples 13900 joint_samples 40 [1045576, 630565] processed_samples 13900 unjoint_samples 13900 joint_samples 41 [1040737, 379854] [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure processed_samples 13901 unjoint_samples 13900 joint_samples 42 [382596, 1043690] processed_samples 13901 unjoint_samples 13900 joint_samples 41 [472711, 1046881] processed_samples 13901 unjoint_samples 13900 joint_samples 41 [472711, 1046881] processed_samples 13901 unjoint_samples 13900 joint_samples 41 [300732, 1047078] processed_samples 13901 unjoint_samples 13900 joint_samples 42 [382596, 1043690] processed_samples 13901 unjoint_samples 13900 joint_samples 41 [300732, 1047078] processed_samples 13901 unjoint_samples 13900 joint_samples 42 [1035404, 726089] processed_samples 13902 unjoint_samples 13900 joint_samples 41 [1044163, 311378] processed_samples 13902 unjoint_samples 13900 joint_samples 41 [1044163, 311378] processed_samples 13901 unjoint_samples 13900 joint_samples 42 [1035404, 726089] processed_samples 13901 unjoint_samples 13900 joint_samples 40 [1046639, 780867] processed_samples 13901 unjoint_samples 13900 joint_samples 40 [1046639, 780867] [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure processed_samples 14001 unjoint_samples 14000 joint_samples 41 [1046639, 54343] processed_samples 14001 unjoint_samples 14000 joint_samples 43 [86210, 1039762] processed_samples 14000 unjoint_samples 14000 joint_samples 40 [1045576, 958082] processed_samples 14001 unjoint_samples 14000 joint_samples 41 [902756, 1046881] processed_samples 14001 unjoint_samples 14000 joint_samples 42 [732873, 1043690] processed_samples 14001 unjoint_samples 14000 joint_samples 41 [634779, 1047078] processed_samples 14000 unjoint_samples 14000 joint_samples 41 [1040737, 691165] processed_samples 14002 unjoint_samples 14000 joint_samples 41 [1044163, 802550] processed_samples 14001 unjoint_samples 14000 joint_samples 41 [1046639, 54343] processed_samples 14001 unjoint_samples 14000 joint_samples 43 [86210, 1039762] processed_samples 14000 unjoint_samples 14000 joint_samples 40 [1045576, 958082] processed_samples 14001 unjoint_samples 14000 joint_samples 41 [634779, 1047078] processed_samples 14001 unjoint_samples 14000 joint_samples 42 [732873, 1043690] processed_samples 14000 unjoint_samples 14000 joint_samples 41 [1040737, 691165] processed_samples 14002 unjoint_samples 14000 joint_samples 41 [1044163, 802550] processed_samples 14001 unjoint_samples 14000 joint_samples 41 [902756, 1046881] [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d5ebea00] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure processed_samples 14100 unjoint_samples 14100 joint_samples 41 [317803, 1027078] processed_samples 14100 unjoint_samples 14100 joint_samples 41 [317803, 1027078] processed_samples 14101 unjoint_samples 14100 joint_samples 41 [1046639, 418527] processed_samples 14101 unjoint_samples 14100 joint_samples 41 [1046639, 418527] processed_samples 14101 unjoint_samples 14100 joint_samples 42 [1023502, 130137] processed_samples 14101 unjoint_samples 14100 joint_samples 42 [1023502, 130137] processed_samples 14101 unjoint_samples 14100 joint_samples 43 [356430, 1039762] processed_samples 14101 unjoint_samples 14100 joint_samples 43 [356430, 1039762] processed_samples 14100 unjoint_samples 14100 joint_samples 41 [1040737, 987505] processed_samples 14101 unjoint_samples 14100 joint_samples 42 [1016141, 1043690] [h264 @ 0x561ea4492980] mmco: unref short failure processed_samples 14101 unjoint_samples 14100 joint_samples 41 [1030009, 1047078] processed_samples 14101 unjoint_samples 14100 joint_samples 41 [1030009, 1047078] processed_samples 14101 unjoint_samples 14100 joint_samples 42 [1016141, 1043690] processed_samples 14100 unjoint_samples 14100 joint_samples 41 [1040737, 987505] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure processed_samples 14102 unjoint_samples 14100 joint_samples 42 [1046344, 51661] processed_samples 14102 unjoint_samples 14100 joint_samples 42 [1046344, 51661] [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6d5ebea00] mmco: unref short failure [h264 @ 0x55e6d5ebea00] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure processed_samples 14200 unjoint_samples 14200 joint_samples 41 [581163, 1027078] processed_samples 14200 unjoint_samples 14200 joint_samples 41 [581163, 1027078] processed_samples 14200 unjoint_samples 14200 joint_samples 42 [1046667, 197041] processed_samples 14200 unjoint_samples 14200 joint_samples 42 [1046667, 197041] [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 14201 unjoint_samples 14200 joint_samples 42 [1034583, 297615] processed_samples 14201 unjoint_samples 14200 joint_samples 42 [1034583, 297615] processed_samples 14201 unjoint_samples 14200 joint_samples 43 [1045071, 240822] processed_samples 14201 unjoint_samples 14200 joint_samples 43 [1045071, 240822] processed_samples 14201 unjoint_samples 14200 joint_samples 43 [773028, 1039762] processed_samples 14201 unjoint_samples 14200 joint_samples 43 [773028, 1039762] processed_samples 14201 unjoint_samples 14200 joint_samples 42 [1023502, 445326] processed_samples 14201 unjoint_samples 14200 joint_samples 42 [1023502, 445326] [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure processed_samples 14202 unjoint_samples 14200 joint_samples 42 [1046344, 468044] processed_samples 14202 unjoint_samples 14200 joint_samples 42 [1046344, 468044] processed_samples 14201 unjoint_samples 14200 joint_samples 41 [1046639, 693207] processed_samples 14201 unjoint_samples 14200 joint_samples 41 [1046639, 693207] [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea3a4af00] co located POCs unavailable [h264 @ 0x561ea3a4af00] co located POCs unavailable [h264 @ 0x561ea3a4af00] co located POCs unavailable [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure processed_samples 14300 unjoint_samples 14300 joint_samples 42 [1046667, 473143] [h264 @ 0x55e6d4750cc0] co located POCs unavailable [h264 @ 0x55e6d4750cc0] co located POCs unavailable [h264 @ 0x55e6d4750cc0] co located POCs unavailable processed_samples 14300 unjoint_samples 14300 joint_samples 41 [907789, 1027078] processed_samples 14301 unjoint_samples 14300 joint_samples 42 [1023502, 774782] processed_samples 14301 unjoint_samples 14300 joint_samples 42 [1034583, 602485] processed_samples 14301 unjoint_samples 14300 joint_samples 41 [1046639, 1009297] processed_samples 14301 unjoint_samples 14300 joint_samples 43 [1045071, 488285] processed_samples 14302 unjoint_samples 14300 joint_samples 42 [1046344, 711016] processed_samples 14301 unjoint_samples 14300 joint_samples 43 [1047333, 1046325] processed_samples 14300 unjoint_samples 14300 joint_samples 42 [1046667, 473143] processed_samples 14300 unjoint_samples 14300 joint_samples 41 [907789, 1027078] processed_samples 14301 unjoint_samples 14300 joint_samples 42 [1034583, 602485] processed_samples 14301 unjoint_samples 14300 joint_samples 42 [1023502, 774782] processed_samples 14301 unjoint_samples 14300 joint_samples 41 [1046639, 1009297] processed_samples 14301 unjoint_samples 14300 joint_samples 43 [1045071, 488285] processed_samples 14302 unjoint_samples 14300 joint_samples 42 [1046344, 711016] processed_samples 14301 unjoint_samples 14300 joint_samples 43 [1047333, 1046325] [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure processed_samples 14400 unjoint_samples 14400 joint_samples 42 [216748, 1043976] processed_samples 14400 unjoint_samples 14400 joint_samples 42 [216748, 1043976] [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure processed_samples 14400 unjoint_samples 14400 joint_samples 42 [1046667, 832934] [h264 @ 0x561ea548af40] mmco: unref short failure processed_samples 14401 unjoint_samples 14400 joint_samples 43 [51475, 1041209] processed_samples 14400 unjoint_samples 14400 joint_samples 42 [1046667, 832934] [h264 @ 0x55e6cf8a8940] mmco: unref short failure processed_samples 14401 unjoint_samples 14400 joint_samples 43 [51475, 1041209] processed_samples 14401 unjoint_samples 14400 joint_samples 44 [353266, 1046325] processed_samples 14401 unjoint_samples 14400 joint_samples 42 [442632, 1034070] [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure processed_samples 14401 unjoint_samples 14400 joint_samples 44 [353266, 1046325] processed_samples 14401 unjoint_samples 14400 joint_samples 42 [442632, 1034070] processed_samples 14402 unjoint_samples 14400 joint_samples 42 [1046344, 967506] processed_samples 14401 unjoint_samples 14400 joint_samples 43 [1045071, 841473] processed_samples 14401 unjoint_samples 14400 joint_samples 43 [1045071, 841473] processed_samples 14401 unjoint_samples 14400 joint_samples 42 [1034583, 1034623] processed_samples 14402 unjoint_samples 14400 joint_samples 42 [1046344, 967506] processed_samples 14401 unjoint_samples 14400 joint_samples 42 [1034583, 1034623] [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure processed_samples 14500 unjoint_samples 14500 joint_samples 42 [567691, 1043976] processed_samples 14500 unjoint_samples 14500 joint_samples 42 [567691, 1043976] [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure processed_samples 14501 unjoint_samples 14500 joint_samples 43 [1047946, 313695] processed_samples 14500 unjoint_samples 14500 joint_samples 43 [1047066, 103657] processed_samples 14500 unjoint_samples 14500 joint_samples 43 [1047066, 103657] processed_samples 14501 unjoint_samples 14500 joint_samples 43 [1047946, 313695] processed_samples 14501 unjoint_samples 14500 joint_samples 44 [172435, 1047502] processed_samples 14501 unjoint_samples 14500 joint_samples 43 [363956, 1041209] processed_samples 14501 unjoint_samples 14500 joint_samples 42 [798953, 1034070] processed_samples 14501 unjoint_samples 14500 joint_samples 44 [172435, 1047502] processed_samples 14501 unjoint_samples 14500 joint_samples 42 [798953, 1034070] processed_samples 14501 unjoint_samples 14500 joint_samples 43 [363956, 1041209] [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure processed_samples 14501 unjoint_samples 14500 joint_samples 44 [755149, 1046325] processed_samples 14502 unjoint_samples 14500 joint_samples 43 [1046344, 242634] processed_samples 14501 unjoint_samples 14500 joint_samples 44 [755149, 1046325] processed_samples 14502 unjoint_samples 14500 joint_samples 43 [1046344, 242634] [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d5ebea00] mmco: unref short failure [h264 @ 0x55e6d5ebea00] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6d5ebea00] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea0f3a780] stream 1, offset 0xbcbd: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea0f3a780] stream 1, offset 0xbe30: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6d2168b40] stream 1, offset 0xbcbd: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6d2168b40] stream 1, offset 0xbe30: partial file [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure processed_samples 14600 unjoint_samples 14600 joint_samples 43 [1047066, 492949] processed_samples 14600 unjoint_samples 14600 joint_samples 43 [1047066, 492949] processed_samples 14601 unjoint_samples 14600 joint_samples 43 [735726, 1041209] processed_samples 14601 unjoint_samples 14600 joint_samples 43 [1047627, 46338] processed_samples 14601 unjoint_samples 14600 joint_samples 43 [735726, 1041209] processed_samples 14601 unjoint_samples 14600 joint_samples 43 [1047627, 46338] processed_samples 14601 unjoint_samples 14600 joint_samples 43 [1047946, 637506] processed_samples 14601 unjoint_samples 14600 joint_samples 43 [1047946, 637506] processed_samples 14601 unjoint_samples 14600 joint_samples 44 [451141, 1047502] processed_samples 14601 unjoint_samples 14600 joint_samples 44 [451141, 1047502] processed_samples 14601 unjoint_samples 14600 joint_samples 42 [862530, 1043976] processed_samples 14601 unjoint_samples 14600 joint_samples 42 [862530, 1043976] processed_samples 14602 unjoint_samples 14600 joint_samples 43 [1046344, 513906] processed_samples 14602 unjoint_samples 14600 joint_samples 43 [1046344, 513906] processed_samples 14601 unjoint_samples 14600 joint_samples 44 [1019198, 1046325] processed_samples 14601 unjoint_samples 14600 joint_samples 44 [1019198, 1046325] [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure processed_samples 14701 unjoint_samples 14700 joint_samples 43 [1016257, 182935] processed_samples 14701 unjoint_samples 14700 joint_samples 43 [1016257, 182935] processed_samples 14700 unjoint_samples 14700 joint_samples 43 [1047066, 798829] processed_samples 14700 unjoint_samples 14700 joint_samples 43 [1047066, 798829] processed_samples 14701 unjoint_samples 14700 joint_samples 44 [759125, 1047502] processed_samples 14701 unjoint_samples 14700 joint_samples 44 [759125, 1047502] processed_samples 14701 unjoint_samples 14700 joint_samples 43 [1047627, 418280] processed_samples 14701 unjoint_samples 14700 joint_samples 43 [1047627, 418280] processed_samples 14701 unjoint_samples 14700 joint_samples 44 [180252, 987939] processed_samples 14701 unjoint_samples 14700 joint_samples 44 [180252, 987939] processed_samples 14701 unjoint_samples 14700 joint_samples 45 [227148, 1048307] processed_samples 14701 unjoint_samples 14700 joint_samples 45 [227148, 1048307] [h264 @ 0x561ea46b2340] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 14702 unjoint_samples 14700 joint_samples 43 [1046344, 846463] processed_samples 14702 unjoint_samples 14700 joint_samples 43 [1046344, 846463] processed_samples 14701 unjoint_samples 14700 joint_samples 43 [991882, 1041209] processed_samples 14701 unjoint_samples 14700 joint_samples 43 [991882, 1041209] [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure processed_samples 14800 unjoint_samples 14800 joint_samples 44 [1047332, 58114] processed_samples 14800 unjoint_samples 14800 joint_samples 44 [1047332, 58114] processed_samples 14801 unjoint_samples 14800 joint_samples 44 [1046250, 204752] processed_samples 14801 unjoint_samples 14800 joint_samples 44 [1046250, 204752] processed_samples 14801 unjoint_samples 14800 joint_samples 43 [1016257, 480615] processed_samples 14801 unjoint_samples 14800 joint_samples 43 [1016257, 480615] processed_samples 14801 unjoint_samples 14800 joint_samples 44 [631713, 987939] processed_samples 14801 unjoint_samples 14800 joint_samples 44 [631713, 987939] [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure processed_samples 14802 unjoint_samples 14800 joint_samples 44 [1046344, 68968] processed_samples 14802 unjoint_samples 14800 joint_samples 44 [1046344, 68968] [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure processed_samples 14801 unjoint_samples 14800 joint_samples 45 [543669, 1048307] processed_samples 14801 unjoint_samples 14800 joint_samples 45 [543669, 1048307] processed_samples 14801 unjoint_samples 14800 joint_samples 43 [1047627, 799414] processed_samples 14801 unjoint_samples 14800 joint_samples 43 [1047627, 799414] processed_samples 14801 unjoint_samples 14800 joint_samples 44 [1036038, 1047502] processed_samples 14801 unjoint_samples 14800 joint_samples 44 [1036038, 1047502] [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure processed_samples 14900 unjoint_samples 14900 joint_samples 44 [1047332, 370213] processed_samples 14900 unjoint_samples 14900 joint_samples 44 [1047332, 370213] processed_samples 14901 unjoint_samples 14900 joint_samples 44 [20772, 1046801] processed_samples 14901 unjoint_samples 14900 joint_samples 44 [20772, 1046801] processed_samples 14901 unjoint_samples 14900 joint_samples 44 [1046250, 493120] processed_samples 14901 unjoint_samples 14900 joint_samples 44 [1046250, 493120] processed_samples 14901 unjoint_samples 14900 joint_samples 43 [1016257, 722829] processed_samples 14901 unjoint_samples 14900 joint_samples 43 [1016257, 722829] [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure processed_samples 14901 unjoint_samples 14900 joint_samples 45 [1046612, 317555] processed_samples 14901 unjoint_samples 14900 joint_samples 45 [1046612, 317555] processed_samples 14902 unjoint_samples 14900 joint_samples 44 [1046344, 436620] processed_samples 14901 unjoint_samples 14900 joint_samples 44 [1018645, 1007710] processed_samples 14902 unjoint_samples 14900 joint_samples 44 [1046344, 436620] processed_samples 14901 unjoint_samples 14900 joint_samples 44 [1018645, 1007710] processed_samples 14901 unjoint_samples 14900 joint_samples 45 [826949, 1048307] processed_samples 14901 unjoint_samples 14900 joint_samples 45 [826949, 1048307] [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure processed_samples 15000 unjoint_samples 15000 joint_samples 44 [1047332, 675941] processed_samples 15001 unjoint_samples 15000 joint_samples 45 [199542, 1047133] processed_samples 15001 unjoint_samples 15000 joint_samples 46 [1046971, 176273] processed_samples 15001 unjoint_samples 15000 joint_samples 44 [260454, 1046801] processed_samples 15000 unjoint_samples 15000 joint_samples 44 [1047332, 675941] processed_samples 15001 unjoint_samples 15000 joint_samples 45 [1046612, 587223] processed_samples 15001 unjoint_samples 15000 joint_samples 45 [199542, 1047133] processed_samples 15001 unjoint_samples 15000 joint_samples 46 [1046971, 176273] processed_samples 15001 unjoint_samples 15000 joint_samples 44 [260454, 1046801] processed_samples 15002 unjoint_samples 15000 joint_samples 44 [1046344, 764752] processed_samples 15001 unjoint_samples 15000 joint_samples 45 [1046612, 587223] processed_samples 15001 unjoint_samples 15000 joint_samples 44 [1046250, 951906] processed_samples 15001 unjoint_samples 15000 joint_samples 43 [1016257, 1014472] processed_samples 15001 unjoint_samples 15000 joint_samples 43 [1016257, 1014472] [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure processed_samples 15002 unjoint_samples 15000 joint_samples 44 [1046344, 764752] processed_samples 15001 unjoint_samples 15000 joint_samples 44 [1046250, 951906] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure processed_samples 15100 unjoint_samples 15100 joint_samples 44 [1047332, 1007073] processed_samples 15101 unjoint_samples 15100 joint_samples 45 [261590, 1029946] processed_samples 15101 unjoint_samples 15100 joint_samples 45 [261590, 1029946] processed_samples 15100 unjoint_samples 15100 joint_samples 44 [1047332, 1007073] processed_samples 15101 unjoint_samples 15100 joint_samples 45 [563940, 1047133] processed_samples 15101 unjoint_samples 15100 joint_samples 45 [563940, 1047133] processed_samples 15101 unjoint_samples 15100 joint_samples 44 [581408, 1046801] processed_samples 15101 unjoint_samples 15100 joint_samples 44 [581408, 1046801] processed_samples 15101 unjoint_samples 15100 joint_samples 46 [1046971, 540665] processed_samples 15101 unjoint_samples 15100 joint_samples 46 [1046971, 540665] processed_samples 15101 unjoint_samples 15100 joint_samples 44 [333900, 1046709] processed_samples 15101 unjoint_samples 15100 joint_samples 44 [333900, 1046709] processed_samples 15102 unjoint_samples 15100 joint_samples 45 [93898, 1041499] processed_samples 15102 unjoint_samples 15100 joint_samples 45 [93898, 1041499] processed_samples 15101 unjoint_samples 15100 joint_samples 45 [1046612, 893290] processed_samples 15101 unjoint_samples 15100 joint_samples 45 [1046612, 893290] [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure processed_samples 15200 unjoint_samples 15200 joint_samples 45 [1047332, 198973] [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure processed_samples 15201 unjoint_samples 15200 joint_samples 44 [687954, 1046709] processed_samples 15201 unjoint_samples 15200 joint_samples 44 [1032845, 1046801] processed_samples 15201 unjoint_samples 15200 joint_samples 46 [106283, 1042345] processed_samples 15201 unjoint_samples 15200 joint_samples 46 [1046971, 757401] processed_samples 15202 unjoint_samples 15200 joint_samples 45 [391713, 1041499] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure processed_samples 15200 unjoint_samples 15200 joint_samples 45 [1047332, 198973] processed_samples 15201 unjoint_samples 15200 joint_samples 45 [817513, 1047133] [h264 @ 0x561ea05da800] mmco: unref short failure processed_samples 15201 unjoint_samples 15200 joint_samples 45 [717649, 1029946] [h264 @ 0x55e6d5b9bf40] mmco: unref short failure processed_samples 15201 unjoint_samples 15200 joint_samples 46 [106283, 1042345] processed_samples 15201 unjoint_samples 15200 joint_samples 44 [687954, 1046709] processed_samples 15201 unjoint_samples 15200 joint_samples 45 [817513, 1047133] processed_samples 15201 unjoint_samples 15200 joint_samples 44 [1032845, 1046801] processed_samples 15202 unjoint_samples 15200 joint_samples 45 [391713, 1041499] processed_samples 15201 unjoint_samples 15200 joint_samples 46 [1046971, 757401] processed_samples 15201 unjoint_samples 15200 joint_samples 45 [717649, 1029946] [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 15300 unjoint_samples 15300 joint_samples 45 [1047332, 590429] processed_samples 15301 unjoint_samples 15300 joint_samples 47 [114520, 1003652] processed_samples 15301 unjoint_samples 15300 joint_samples 45 [379627, 1046801] processed_samples 15301 unjoint_samples 15300 joint_samples 46 [336280, 1042345] processed_samples 15301 unjoint_samples 15300 joint_samples 44 [918247, 1046709] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure processed_samples 15301 unjoint_samples 15300 joint_samples 45 [990094, 1029946] [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 15300 unjoint_samples 15300 joint_samples 45 [1047332, 590429] processed_samples 15301 unjoint_samples 15300 joint_samples 47 [114520, 1003652] processed_samples 15301 unjoint_samples 15300 joint_samples 45 [379627, 1046801] processed_samples 15301 unjoint_samples 15300 joint_samples 46 [67675, 1047133] processed_samples 15302 unjoint_samples 15300 joint_samples 45 [712318, 1041499] processed_samples 15301 unjoint_samples 15300 joint_samples 46 [336280, 1042345] processed_samples 15301 unjoint_samples 15300 joint_samples 44 [918247, 1046709] [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure processed_samples 15301 unjoint_samples 15300 joint_samples 45 [990094, 1029946] processed_samples 15301 unjoint_samples 15300 joint_samples 46 [67675, 1047133] [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure processed_samples 15302 unjoint_samples 15300 joint_samples 45 [712318, 1041499] [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure processed_samples 15400 unjoint_samples 15400 joint_samples 45 [1047332, 939850] processed_samples 15400 unjoint_samples 15400 joint_samples 45 [1047332, 939850] processed_samples 15401 unjoint_samples 15400 joint_samples 46 [351976, 1047133] processed_samples 15401 unjoint_samples 15400 joint_samples 45 [1029500, 202305] processed_samples 15401 unjoint_samples 15400 joint_samples 47 [438043, 1003652] processed_samples 15401 unjoint_samples 15400 joint_samples 46 [351976, 1047133] processed_samples 15401 unjoint_samples 15400 joint_samples 46 [234701, 1047159] processed_samples 15401 unjoint_samples 15400 joint_samples 45 [1029500, 202305] processed_samples 15401 unjoint_samples 15400 joint_samples 46 [234701, 1047159] processed_samples 15401 unjoint_samples 15400 joint_samples 47 [438043, 1003652] processed_samples 15401 unjoint_samples 15400 joint_samples 45 [832618, 1046801] processed_samples 15401 unjoint_samples 15400 joint_samples 45 [832618, 1046801] processed_samples 15401 unjoint_samples 15400 joint_samples 46 [725440, 1042345] processed_samples 15401 unjoint_samples 15400 joint_samples 46 [725440, 1042345] processed_samples 15402 unjoint_samples 15400 joint_samples 45 [949369, 1041499] [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure processed_samples 15402 unjoint_samples 15400 joint_samples 45 [949369, 1041499] [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure processed_samples 15500 unjoint_samples 15500 joint_samples 46 [415447, 989525] processed_samples 15500 unjoint_samples 15500 joint_samples 46 [415447, 989525] [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure processed_samples 15501 unjoint_samples 15500 joint_samples 45 [1029500, 560494] processed_samples 15501 unjoint_samples 15500 joint_samples 47 [986142, 110488] processed_samples 15501 unjoint_samples 15500 joint_samples 47 [986142, 110488] processed_samples 15501 unjoint_samples 15500 joint_samples 45 [1029500, 560494] processed_samples 15501 unjoint_samples 15500 joint_samples 46 [89622, 1046801] processed_samples 15501 unjoint_samples 15500 joint_samples 46 [89622, 1046801] processed_samples 15501 unjoint_samples 15500 joint_samples 47 [717969, 1003652] processed_samples 15501 unjoint_samples 15500 joint_samples 46 [702568, 1047133] processed_samples 15501 unjoint_samples 15500 joint_samples 47 [717969, 1003652] processed_samples 15501 unjoint_samples 15500 joint_samples 46 [702568, 1047133] processed_samples 15502 unjoint_samples 15500 joint_samples 46 [131255, 1044972] processed_samples 15502 unjoint_samples 15500 joint_samples 46 [131255, 1044972] processed_samples 15501 unjoint_samples 15500 joint_samples 46 [581195, 1047159] processed_samples 15501 unjoint_samples 15500 joint_samples 46 [581195, 1047159] [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure processed_samples 15600 unjoint_samples 15600 joint_samples 46 [831073, 989525] processed_samples 15600 unjoint_samples 15600 joint_samples 46 [831073, 989525] processed_samples 15601 unjoint_samples 15600 joint_samples 47 [1043206, 79672] processed_samples 15601 unjoint_samples 15600 joint_samples 46 [448518, 1046801] processed_samples 15601 unjoint_samples 15600 joint_samples 47 [1043206, 79672] processed_samples 15601 unjoint_samples 15600 joint_samples 46 [927355, 1047159] processed_samples 15601 unjoint_samples 15600 joint_samples 47 [986142, 358805] processed_samples 15601 unjoint_samples 15600 joint_samples 47 [1024467, 1026364] processed_samples 15601 unjoint_samples 15600 joint_samples 46 [448518, 1046801] processed_samples 15602 unjoint_samples 15600 joint_samples 46 [403950, 1044972] processed_samples 15601 unjoint_samples 15600 joint_samples 46 [927355, 1047159] processed_samples 15601 unjoint_samples 15600 joint_samples 47 [986142, 358805] processed_samples 15601 unjoint_samples 15600 joint_samples 47 [1024467, 1026364] processed_samples 15602 unjoint_samples 15600 joint_samples 46 [403950, 1044972] processed_samples 15601 unjoint_samples 15600 joint_samples 45 [1029500, 869117] processed_samples 15601 unjoint_samples 15600 joint_samples 45 [1029500, 869117] [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d1ab3a00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1ab3a00] mmco: unref short failure [h264 @ 0x55e6d1ab3a00] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea1f53700] mmco: unref short failure [h264 @ 0x561ea1f53700] mmco: unref short failure processed_samples 15700 unjoint_samples 15700 joint_samples 47 [1046281, 59784] processed_samples 15700 unjoint_samples 15700 joint_samples 47 [1046281, 59784] processed_samples 15701 unjoint_samples 15700 joint_samples 47 [1043206, 344694] processed_samples 15701 unjoint_samples 15700 joint_samples 48 [1043022, 288206] processed_samples 15701 unjoint_samples 15700 joint_samples 47 [1043206, 344694] processed_samples 15701 unjoint_samples 15700 joint_samples 48 [1043022, 288206] processed_samples 15701 unjoint_samples 15700 joint_samples 47 [94454, 1047159] processed_samples 15701 unjoint_samples 15700 joint_samples 47 [94454, 1047159] processed_samples 15701 unjoint_samples 15700 joint_samples 46 [1043076, 128869] processed_samples 15701 unjoint_samples 15700 joint_samples 46 [1043076, 128869] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure processed_samples 15701 unjoint_samples 15700 joint_samples 46 [948130, 1046801] processed_samples 15701 unjoint_samples 15700 joint_samples 46 [948130, 1046801] processed_samples 15702 unjoint_samples 15700 joint_samples 46 [742098, 1044972] processed_samples 15702 unjoint_samples 15700 joint_samples 46 [742098, 1044972] processed_samples 15701 unjoint_samples 15700 joint_samples 47 [986142, 611158] processed_samples 15701 unjoint_samples 15700 joint_samples 47 [986142, 611158] [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure processed_samples 15800 unjoint_samples 15800 joint_samples 47 [1046281, 449982] processed_samples 15801 unjoint_samples 15800 joint_samples 46 [1043076, 432438] processed_samples 15801 unjoint_samples 15800 joint_samples 47 [1020791, 293517] [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 15800 unjoint_samples 15800 joint_samples 47 [1046281, 449982] processed_samples 15801 unjoint_samples 15800 joint_samples 48 [1043022, 546617] processed_samples 15801 unjoint_samples 15800 joint_samples 46 [1043076, 432438] processed_samples 15801 unjoint_samples 15800 joint_samples 47 [1020791, 293517] processed_samples 15801 unjoint_samples 15800 joint_samples 47 [498669, 1047159] [h264 @ 0x55e6d5b9bf40] mmco: unref short failure processed_samples 15801 unjoint_samples 15800 joint_samples 47 [498669, 1047159] processed_samples 15801 unjoint_samples 15800 joint_samples 48 [1043022, 546617] [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 15801 unjoint_samples 15800 joint_samples 47 [986142, 923564] processed_samples 15801 unjoint_samples 15800 joint_samples 47 [986142, 923564] processed_samples 15801 unjoint_samples 15800 joint_samples 47 [1043206, 594811] processed_samples 15801 unjoint_samples 15800 joint_samples 47 [1043206, 594811] processed_samples 15802 unjoint_samples 15800 joint_samples 46 [1002249, 1044972] processed_samples 15802 unjoint_samples 15800 joint_samples 46 [1002249, 1044972] [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d1ab3a00] mmco: unref short failure [h264 @ 0x55e6d1ab3a00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure processed_samples 15900 unjoint_samples 15900 joint_samples 47 [1046281, 793804] processed_samples 15900 unjoint_samples 15900 joint_samples 47 [1046281, 793804] processed_samples 15901 unjoint_samples 15900 joint_samples 47 [1020791, 624656] processed_samples 15901 unjoint_samples 15900 joint_samples 47 [1020791, 624656] processed_samples 15901 unjoint_samples 15900 joint_samples 47 [916475, 1047159] processed_samples 15901 unjoint_samples 15900 joint_samples 47 [916475, 1047159] processed_samples 15901 unjoint_samples 15900 joint_samples 48 [1042412, 72493] processed_samples 15901 unjoint_samples 15900 joint_samples 48 [1043022, 830371] processed_samples 15901 unjoint_samples 15900 joint_samples 46 [1043076, 733003] processed_samples 15901 unjoint_samples 15900 joint_samples 46 [1043076, 733003] processed_samples 15901 unjoint_samples 15900 joint_samples 48 [1042412, 72493] processed_samples 15902 unjoint_samples 15900 joint_samples 47 [323195, 1047351] processed_samples 15902 unjoint_samples 15900 joint_samples 47 [323195, 1047351] processed_samples 15901 unjoint_samples 15900 joint_samples 48 [1043022, 830371] processed_samples 15901 unjoint_samples 15900 joint_samples 47 [1043206, 837844] processed_samples 15901 unjoint_samples 15900 joint_samples 47 [1043206, 837844] [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 16000 unjoint_samples 16000 joint_samples 48 [1046281, 14612] processed_samples 16000 unjoint_samples 16000 joint_samples 48 [1046281, 14612] processed_samples 16001 unjoint_samples 16000 joint_samples 48 [1042412, 474231] processed_samples 16001 unjoint_samples 16000 joint_samples 48 [1042412, 474231] processed_samples 16001 unjoint_samples 16000 joint_samples 48 [1027820, 208127] processed_samples 16001 unjoint_samples 16000 joint_samples 48 [1027820, 208127] processed_samples 16001 unjoint_samples 16000 joint_samples 49 [22076, 1046638] processed_samples 16001 unjoint_samples 16000 joint_samples 49 [22076, 1046638] processed_samples 16001 unjoint_samples 16000 joint_samples 48 [1045500, 31312] processed_samples 16001 unjoint_samples 16000 joint_samples 48 [1045500, 31312] processed_samples 16001 unjoint_samples 16000 joint_samples 47 [1020791, 983134] processed_samples 16001 unjoint_samples 16000 joint_samples 47 [1020791, 983134] processed_samples 16002 unjoint_samples 16000 joint_samples 47 [816348, 1047351] processed_samples 16002 unjoint_samples 16000 joint_samples 47 [816348, 1047351] processed_samples 16001 unjoint_samples 16000 joint_samples 46 [1043076, 987437] processed_samples 16001 unjoint_samples 16000 joint_samples 46 [1043076, 987437] [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea5f27180] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure processed_samples 16100 unjoint_samples 16100 joint_samples 48 [1046281, 363624] processed_samples 16100 unjoint_samples 16100 joint_samples 48 [1046281, 363624] processed_samples 16101 unjoint_samples 16100 joint_samples 47 [1048044, 202030] processed_samples 16101 unjoint_samples 16100 joint_samples 47 [1048044, 202030] processed_samples 16101 unjoint_samples 16100 joint_samples 49 [396263, 1046638] processed_samples 16101 unjoint_samples 16100 joint_samples 49 [396263, 1046638] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1027820, 534123] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1027820, 534123] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1027119, 325014] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1027119, 325014] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1045500, 324517] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1045500, 324517] processed_samples 16102 unjoint_samples 16100 joint_samples 48 [1046952, 138635] processed_samples 16102 unjoint_samples 16100 joint_samples 48 [1046952, 138635] [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1042412, 890807] processed_samples 16101 unjoint_samples 16100 joint_samples 48 [1042412, 890807] [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 16201 unjoint_samples 16200 joint_samples 49 [669750, 1046638] processed_samples 16200 unjoint_samples 16200 joint_samples 48 [1046281, 690092] processed_samples 16201 unjoint_samples 16200 joint_samples 49 [669750, 1046638] processed_samples 16201 unjoint_samples 16200 joint_samples 49 [1046765, 149593] processed_samples 16200 unjoint_samples 16200 joint_samples 48 [1046281, 690092] processed_samples 16201 unjoint_samples 16200 joint_samples 49 [1046765, 149593] processed_samples 16201 unjoint_samples 16200 joint_samples 48 [1027119, 626171] processed_samples 16201 unjoint_samples 16200 joint_samples 48 [1027119, 626171] processed_samples 16201 unjoint_samples 16200 joint_samples 48 [1027820, 812692] processed_samples 16201 unjoint_samples 16200 joint_samples 48 [1027820, 812692] processed_samples 16201 unjoint_samples 16200 joint_samples 47 [1048044, 580734] [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure processed_samples 16202 unjoint_samples 16200 joint_samples 48 [1046952, 444828] processed_samples 16201 unjoint_samples 16200 joint_samples 47 [1048044, 580734] [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 16202 unjoint_samples 16200 joint_samples 48 [1046952, 444828] processed_samples 16201 unjoint_samples 16200 joint_samples 48 [1045500, 716147] processed_samples 16201 unjoint_samples 16200 joint_samples 48 [1045500, 716147] [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce544200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure processed_samples 16301 unjoint_samples 16300 joint_samples 48 [1027119, 941842] processed_samples 16301 unjoint_samples 16300 joint_samples 48 [1027119, 941842] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [1047532, 12059] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [1047532, 12059] processed_samples 16301 unjoint_samples 16300 joint_samples 47 [1048044, 858929] processed_samples 16301 unjoint_samples 16300 joint_samples 47 [1048044, 858929] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [188572, 988726] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [188572, 988726] processed_samples 16300 unjoint_samples 16300 joint_samples 48 [1046281, 991576] processed_samples 16300 unjoint_samples 16300 joint_samples 48 [1046281, 991576] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [1046765, 447956] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [1046765, 447956] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [995661, 1046638] processed_samples 16301 unjoint_samples 16300 joint_samples 49 [995661, 1046638] processed_samples 16302 unjoint_samples 16300 joint_samples 48 [1046952, 750652] processed_samples 16302 unjoint_samples 16300 joint_samples 48 [1046952, 750652] [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure processed_samples 16400 unjoint_samples 16400 joint_samples 49 [279906, 1048162] processed_samples 16400 unjoint_samples 16400 joint_samples 49 [279906, 1048162] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [269073, 1030233] processed_samples 16401 unjoint_samples 16400 joint_samples 50 [198666, 1046638] processed_samples 16401 unjoint_samples 16400 joint_samples 48 [113959, 1047589] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [499384, 988726] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [1047532, 229616] processed_samples 16401 unjoint_samples 16400 joint_samples 50 [198666, 1046638] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [269073, 1030233] processed_samples 16401 unjoint_samples 16400 joint_samples 48 [113959, 1047589] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [1046765, 806486] [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 16401 unjoint_samples 16400 joint_samples 49 [499384, 988726] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [1047532, 229616] processed_samples 16401 unjoint_samples 16400 joint_samples 49 [1046765, 806486] [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure processed_samples 16402 unjoint_samples 16400 joint_samples 48 [1046952, 1016357] processed_samples 16402 unjoint_samples 16400 joint_samples 48 [1046952, 1016357] [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 16501 unjoint_samples 16500 joint_samples 50 [24749, 1037804] processed_samples 16501 unjoint_samples 16500 joint_samples 50 [24749, 1037804] processed_samples 16500 unjoint_samples 16500 joint_samples 49 [532311, 1048162] processed_samples 16500 unjoint_samples 16500 joint_samples 49 [532311, 1048162] processed_samples 16501 unjoint_samples 16500 joint_samples 48 [426475, 1047589] processed_samples 16501 unjoint_samples 16500 joint_samples 48 [426475, 1047589] processed_samples 16501 unjoint_samples 16500 joint_samples 49 [1047532, 461636] processed_samples 16501 unjoint_samples 16500 joint_samples 50 [527850, 1046638] processed_samples 16501 unjoint_samples 16500 joint_samples 49 [1047532, 461636] processed_samples 16501 unjoint_samples 16500 joint_samples 50 [527850, 1046638] processed_samples 16502 unjoint_samples 16500 joint_samples 49 [1046952, 205659] processed_samples 16502 unjoint_samples 16500 joint_samples 49 [1046952, 205659] processed_samples 16501 unjoint_samples 16500 joint_samples 49 [490485, 1030233] processed_samples 16501 unjoint_samples 16500 joint_samples 49 [490485, 1030233] processed_samples 16501 unjoint_samples 16500 joint_samples 49 [793268, 988726] processed_samples 16501 unjoint_samples 16500 joint_samples 49 [793268, 988726] [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure processed_samples 16601 unjoint_samples 16600 joint_samples 48 [778242, 1047589] [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure processed_samples 16601 unjoint_samples 16600 joint_samples 50 [310190, 1037804] processed_samples 16601 unjoint_samples 16600 joint_samples 49 [1047532, 723656] processed_samples 16600 unjoint_samples 16600 joint_samples 49 [839249, 1048162] processed_samples 16601 unjoint_samples 16600 joint_samples 49 [1038283, 1036514] [h264 @ 0x55e6cdce6a40] mmco: unref short failure processed_samples 16602 unjoint_samples 16600 joint_samples 49 [1046952, 631240] processed_samples 16601 unjoint_samples 16600 joint_samples 50 [779961, 1046638] processed_samples 16601 unjoint_samples 16600 joint_samples 49 [762079, 1030233] [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure processed_samples 16600 unjoint_samples 16600 joint_samples 49 [839249, 1048162] processed_samples 16601 unjoint_samples 16600 joint_samples 48 [778242, 1047589] processed_samples 16601 unjoint_samples 16600 joint_samples 50 [310190, 1037804] processed_samples 16601 unjoint_samples 16600 joint_samples 49 [1047532, 723656] processed_samples 16602 unjoint_samples 16600 joint_samples 49 [1046952, 631240] [h264 @ 0x55e6d199ac00] mmco: unref short failure processed_samples 16601 unjoint_samples 16600 joint_samples 49 [762079, 1030233] processed_samples 16601 unjoint_samples 16600 joint_samples 50 [779961, 1046638] processed_samples 16601 unjoint_samples 16600 joint_samples 49 [1038283, 1036514] [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 16700 unjoint_samples 16700 joint_samples 50 [1047971, 18189] processed_samples 16700 unjoint_samples 16700 joint_samples 50 [1047971, 18189] processed_samples 16701 unjoint_samples 16700 joint_samples 50 [397942, 1043461] processed_samples 16701 unjoint_samples 16700 joint_samples 50 [397942, 1043461] processed_samples 16701 unjoint_samples 16700 joint_samples 51 [1045604, 87507] processed_samples 16701 unjoint_samples 16700 joint_samples 51 [1045604, 87507] processed_samples 16701 unjoint_samples 16700 joint_samples 48 [995484, 1047589] processed_samples 16701 unjoint_samples 16700 joint_samples 48 [995484, 1047589] processed_samples 16701 unjoint_samples 16700 joint_samples 49 [976636, 1030233] processed_samples 16701 unjoint_samples 16700 joint_samples 50 [570793, 1037804] processed_samples 16701 unjoint_samples 16700 joint_samples 49 [976636, 1030233] processed_samples 16701 unjoint_samples 16700 joint_samples 50 [101784, 1023532] processed_samples 16701 unjoint_samples 16700 joint_samples 50 [570793, 1037804] processed_samples 16701 unjoint_samples 16700 joint_samples 50 [101784, 1023532] processed_samples 16702 unjoint_samples 16700 joint_samples 49 [1046952, 952260] processed_samples 16702 unjoint_samples 16700 joint_samples 49 [1046952, 952260] [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea06f8980] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure processed_samples 16800 unjoint_samples 16800 joint_samples 50 [1047971, 291868] processed_samples 16800 unjoint_samples 16800 joint_samples 50 [1047971, 291868] processed_samples 16801 unjoint_samples 16800 joint_samples 51 [1045604, 430525] processed_samples 16801 unjoint_samples 16800 joint_samples 49 [1047315, 187107] processed_samples 16801 unjoint_samples 16800 joint_samples 49 [1047315, 187107] processed_samples 16801 unjoint_samples 16800 joint_samples 51 [1045604, 430525] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [401003, 1023532] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [401003, 1023532] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [671795, 1043461] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [671795, 1043461] processed_samples 16802 unjoint_samples 16800 joint_samples 50 [294600, 996557] processed_samples 16802 unjoint_samples 16800 joint_samples 50 [294600, 996557] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [959265, 1037804] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [959265, 1037804] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [1046428, 279135] processed_samples 16801 unjoint_samples 16800 joint_samples 50 [1046428, 279135] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1029963, 50075] processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1029963, 50075] [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure processed_samples 16901 unjoint_samples 16900 joint_samples 49 [1047315, 497834] [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure processed_samples 16901 unjoint_samples 16900 joint_samples 49 [1047315, 497834] processed_samples 16901 unjoint_samples 16900 joint_samples 50 [1046428, 560199] processed_samples 16900 unjoint_samples 16900 joint_samples 50 [1047971, 546832] processed_samples 16901 unjoint_samples 16900 joint_samples 50 [1046428, 560199] processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1046975, 133468] processed_samples 16900 unjoint_samples 16900 joint_samples 50 [1047971, 546832] processed_samples 16901 unjoint_samples 16900 joint_samples 50 [700221, 1023532] processed_samples 16901 unjoint_samples 16900 joint_samples 50 [700221, 1023532] processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1046975, 133468] processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1045604, 729904] processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1045604, 729904] processed_samples 16902 unjoint_samples 16900 joint_samples 50 [603959, 996557] processed_samples 16902 unjoint_samples 16900 joint_samples 50 [603959, 996557] [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure processed_samples 17000 unjoint_samples 17000 joint_samples 50 [1047971, 910253] processed_samples 17000 unjoint_samples 17000 joint_samples 50 [1047971, 910253] processed_samples 17001 unjoint_samples 17000 joint_samples 52 [16074, 1047609] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1045243, 128329] processed_samples 17001 unjoint_samples 17000 joint_samples 52 [16074, 1047609] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1045243, 128329] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1029963, 365749] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1029963, 365749] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1046975, 600669] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1046428, 2262] processed_samples 17002 unjoint_samples 17000 joint_samples 50 [943423, 996557] processed_samples 17001 unjoint_samples 17000 joint_samples 49 [1047315, 786246] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1046428, 2262] processed_samples 17001 unjoint_samples 17000 joint_samples 49 [1047315, 786246] processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1046975, 600669] processed_samples 17002 unjoint_samples 17000 joint_samples 50 [943423, 996557] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure processed_samples 17100 unjoint_samples 17100 joint_samples 51 [252391, 1032074] processed_samples 17100 unjoint_samples 17100 joint_samples 51 [252391, 1032074] processed_samples 17101 unjoint_samples 17100 joint_samples 50 [207985, 968095] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1046428, 323724] processed_samples 17101 unjoint_samples 17100 joint_samples 52 [390268, 1047609] processed_samples 17101 unjoint_samples 17100 joint_samples 50 [207985, 968095] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1046428, 323724] processed_samples 17101 unjoint_samples 17100 joint_samples 52 [390268, 1047609] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1029963, 709493] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1029963, 709493] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1046975, 883876] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1046975, 883876] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1045243, 584820] processed_samples 17102 unjoint_samples 17100 joint_samples 51 [1046838, 91216] processed_samples 17101 unjoint_samples 17100 joint_samples 51 [1045243, 584820] processed_samples 17102 unjoint_samples 17100 joint_samples 51 [1046838, 91216] [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6ce2c0c40] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea1f53700] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure processed_samples 17201 unjoint_samples 17200 joint_samples 52 [10372, 1043738] processed_samples 17201 unjoint_samples 17200 joint_samples 52 [10372, 1043738] processed_samples 17201 unjoint_samples 17200 joint_samples 50 [455920, 968095] processed_samples 17201 unjoint_samples 17200 joint_samples 50 [455920, 968095] processed_samples 17201 unjoint_samples 17200 joint_samples 51 [1045243, 916021] processed_samples 17201 unjoint_samples 17200 joint_samples 51 [1045243, 916021] processed_samples 17200 unjoint_samples 17200 joint_samples 51 [654972, 1032074] processed_samples 17200 unjoint_samples 17200 joint_samples 51 [654972, 1032074] processed_samples 17201 unjoint_samples 17200 joint_samples 52 [1046975, 87143] processed_samples 17201 unjoint_samples 17200 joint_samples 52 [1046975, 87143] processed_samples 17202 unjoint_samples 17200 joint_samples 51 [1046838, 371319] processed_samples 17202 unjoint_samples 17200 joint_samples 51 [1046838, 371319] processed_samples 17201 unjoint_samples 17200 joint_samples 51 [1046428, 625101] processed_samples 17201 unjoint_samples 17200 joint_samples 51 [1046428, 625101] processed_samples 17201 unjoint_samples 17200 joint_samples 52 [923624, 1047609] processed_samples 17201 unjoint_samples 17200 joint_samples 52 [923624, 1047609] [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6d5d311c0] stream 0, offset 0x900676: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea94d7900] stream 0, offset 0x900676: partial file [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 17300 unjoint_samples 17300 joint_samples 51 [853323, 1032074] processed_samples 17301 unjoint_samples 17300 joint_samples 53 [1043382, 148715] processed_samples 17301 unjoint_samples 17300 joint_samples 52 [329856, 1043738] processed_samples 17301 unjoint_samples 17300 joint_samples 51 [1046428, 899441] processed_samples 17302 unjoint_samples 17300 joint_samples 52 [1047194, 154209] processed_samples 17301 unjoint_samples 17300 joint_samples 52 [1046975, 456586] processed_samples 17302 unjoint_samples 17300 joint_samples 51 [1046838, 658096] processed_samples 17301 unjoint_samples 17300 joint_samples 50 [813743, 968095] [h264 @ 0x561ea23d1800] mmco: unref short failure processed_samples 17301 unjoint_samples 17300 joint_samples 52 [329856, 1043738] processed_samples 17300 unjoint_samples 17300 joint_samples 51 [853323, 1032074] [h264 @ 0x561ea47f8fc0] mmco: unref short failure processed_samples 17301 unjoint_samples 17300 joint_samples 53 [1043382, 148715] processed_samples 17302 unjoint_samples 17300 joint_samples 52 [1047194, 154209] processed_samples 17301 unjoint_samples 17300 joint_samples 52 [1046975, 456586] processed_samples 17301 unjoint_samples 17300 joint_samples 51 [1046428, 899441] processed_samples 17301 unjoint_samples 17300 joint_samples 50 [813743, 968095] processed_samples 17302 unjoint_samples 17300 joint_samples 51 [1046838, 658096] [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x55e6ce23a1c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cfd581c0] mmco: unref short failure [h264 @ 0x55e6cfd581c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure processed_samples 17401 unjoint_samples 17400 joint_samples 51 [73573, 1046900] processed_samples 17400 unjoint_samples 17400 joint_samples 52 [1046275, 148837] processed_samples 17401 unjoint_samples 17400 joint_samples 53 [1043382, 471724] processed_samples 17401 unjoint_samples 17400 joint_samples 52 [325410, 1030318] processed_samples 17401 unjoint_samples 17400 joint_samples 51 [73573, 1046900] processed_samples 17402 unjoint_samples 17400 joint_samples 52 [1047194, 467163] processed_samples 17400 unjoint_samples 17400 joint_samples 52 [1046275, 148837] processed_samples 17401 unjoint_samples 17400 joint_samples 52 [325410, 1030318] processed_samples 17401 unjoint_samples 17400 joint_samples 53 [1043382, 471724] processed_samples 17401 unjoint_samples 17400 joint_samples 52 [587812, 1043738] processed_samples 17402 unjoint_samples 17400 joint_samples 52 [1047194, 467163] processed_samples 17401 unjoint_samples 17400 joint_samples 52 [587812, 1043738] processed_samples 17402 unjoint_samples 17400 joint_samples 51 [1046838, 974135] processed_samples 17401 unjoint_samples 17400 joint_samples 52 [1046975, 868789] processed_samples 17401 unjoint_samples 17400 joint_samples 52 [1046975, 868789] processed_samples 17402 unjoint_samples 17400 joint_samples 51 [1046838, 974135] [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea11b3180] stream 1, offset 0x14000d8: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6cd7f7a80] stream 1, offset 0x14000d8: partial file [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure processed_samples 17500 unjoint_samples 17500 joint_samples 52 [1046275, 516560] processed_samples 17502 unjoint_samples 17500 joint_samples 51 [422395, 1046900] processed_samples 17501 unjoint_samples 17500 joint_samples 53 [186828, 997468] processed_samples 17500 unjoint_samples 17500 joint_samples 52 [1046275, 516560] processed_samples 17501 unjoint_samples 17500 joint_samples 52 [739452, 1030318] processed_samples 17502 unjoint_samples 17500 joint_samples 51 [422395, 1046900] processed_samples 17501 unjoint_samples 17500 joint_samples 53 [186828, 997468] processed_samples 17501 unjoint_samples 17500 joint_samples 53 [1043382, 835032] processed_samples 17501 unjoint_samples 17500 joint_samples 52 [930796, 1043738] processed_samples 17501 unjoint_samples 17500 joint_samples 52 [739452, 1030318] processed_samples 17501 unjoint_samples 17500 joint_samples 52 [930796, 1043738] processed_samples 17501 unjoint_samples 17500 joint_samples 53 [1043382, 835032] processed_samples 17502 unjoint_samples 17500 joint_samples 52 [1047194, 821917] processed_samples 17502 unjoint_samples 17500 joint_samples 52 [283485, 1039697] processed_samples 17502 unjoint_samples 17500 joint_samples 52 [283485, 1039697] processed_samples 17502 unjoint_samples 17500 joint_samples 52 [1047194, 821917] [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure processed_samples 17601 unjoint_samples 17600 joint_samples 53 [1031640, 293538] processed_samples 17601 unjoint_samples 17600 joint_samples 53 [1031640, 293538] processed_samples 17601 unjoint_samples 17600 joint_samples 54 [84274, 1042309] processed_samples 17601 unjoint_samples 17600 joint_samples 54 [84274, 1042309] processed_samples 17600 unjoint_samples 17600 joint_samples 52 [1046275, 1007788] processed_samples 17600 unjoint_samples 17600 joint_samples 52 [1046275, 1007788] processed_samples 17602 unjoint_samples 17600 joint_samples 53 [1047194, 18942] processed_samples 17602 unjoint_samples 17600 joint_samples 53 [1047194, 18942] processed_samples 17601 unjoint_samples 17600 joint_samples 53 [505482, 997468] processed_samples 17601 unjoint_samples 17600 joint_samples 53 [505482, 997468] processed_samples 17601 unjoint_samples 17600 joint_samples 52 [1008601, 1030318] processed_samples 17601 unjoint_samples 17600 joint_samples 52 [1008601, 1030318] processed_samples 17602 unjoint_samples 17600 joint_samples 51 [711476, 1046900] processed_samples 17602 unjoint_samples 17600 joint_samples 51 [711476, 1046900] [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 17602 unjoint_samples 17600 joint_samples 52 [665263, 1039697] processed_samples 17602 unjoint_samples 17600 joint_samples 52 [665263, 1039697] [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea48d0c80] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure processed_samples 17700 unjoint_samples 17700 joint_samples 53 [1046275, 250233] processed_samples 17700 unjoint_samples 17700 joint_samples 53 [1046275, 250233] processed_samples 17701 unjoint_samples 17700 joint_samples 53 [1031640, 695359] processed_samples 17701 unjoint_samples 17700 joint_samples 53 [1031640, 695359] processed_samples 17701 unjoint_samples 17700 joint_samples 54 [345426, 1042309] processed_samples 17701 unjoint_samples 17700 joint_samples 54 [345426, 1042309] processed_samples 17702 unjoint_samples 17700 joint_samples 53 [1047194, 298885] processed_samples 17702 unjoint_samples 17700 joint_samples 53 [1047969, 17032] processed_samples 17702 unjoint_samples 17700 joint_samples 53 [1047194, 298885] processed_samples 17702 unjoint_samples 17700 joint_samples 53 [1047969, 17032] processed_samples 17701 unjoint_samples 17700 joint_samples 53 [1046645, 277499] processed_samples 17701 unjoint_samples 17700 joint_samples 53 [1046645, 277499] processed_samples 17701 unjoint_samples 17700 joint_samples 53 [734258, 997468] processed_samples 17701 unjoint_samples 17700 joint_samples 53 [734258, 997468] processed_samples 17702 unjoint_samples 17700 joint_samples 51 [1035241, 1046900] processed_samples 17702 unjoint_samples 17700 joint_samples 51 [1035241, 1046900] [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x561ea074af00] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure processed_samples 17801 unjoint_samples 17800 joint_samples 54 [53747, 1039875] processed_samples 17801 unjoint_samples 17800 joint_samples 54 [53747, 1039875] processed_samples 17801 unjoint_samples 17800 joint_samples 53 [1046645, 532959] processed_samples 17800 unjoint_samples 17800 joint_samples 53 [1046275, 648257] processed_samples 17801 unjoint_samples 17800 joint_samples 54 [678561, 1042309] processed_samples 17801 unjoint_samples 17800 joint_samples 53 [1046645, 532959] [h264 @ 0x561ea4e60d40] mmco: unref short failure processed_samples 17800 unjoint_samples 17800 joint_samples 53 [1046275, 648257] processed_samples 17801 unjoint_samples 17800 joint_samples 54 [678561, 1042309] [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure processed_samples 17802 unjoint_samples 17800 joint_samples 53 [1047969, 465372] processed_samples 17802 unjoint_samples 17800 joint_samples 53 [1047969, 465372] [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure processed_samples 17802 unjoint_samples 17800 joint_samples 52 [342649, 1046900] processed_samples 17802 unjoint_samples 17800 joint_samples 52 [342649, 1046900] processed_samples 17802 unjoint_samples 17800 joint_samples 53 [1047194, 728986] processed_samples 17802 unjoint_samples 17800 joint_samples 53 [1047194, 728986] processed_samples 17801 unjoint_samples 17800 joint_samples 53 [1023945, 1022931] processed_samples 17801 unjoint_samples 17800 joint_samples 53 [1023945, 1022931] [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure processed_samples 17901 unjoint_samples 17900 joint_samples 54 [355802, 1039875] processed_samples 17901 unjoint_samples 17900 joint_samples 53 [1046645, 835047] processed_samples 17900 unjoint_samples 17900 joint_samples 53 [1046275, 881372] processed_samples 17901 unjoint_samples 17900 joint_samples 54 [266611, 1046346] processed_samples 17902 unjoint_samples 17900 joint_samples 52 [672743, 1046900] processed_samples 17901 unjoint_samples 17900 joint_samples 54 [355802, 1039875] processed_samples 17902 unjoint_samples 17900 joint_samples 54 [127258, 1046568] processed_samples 17901 unjoint_samples 17900 joint_samples 53 [1046645, 835047] processed_samples 17900 unjoint_samples 17900 joint_samples 53 [1046275, 881372] processed_samples 17901 unjoint_samples 17900 joint_samples 54 [936743, 1042309] [h264 @ 0x55e6cd91c9c0] mmco: unref short failure processed_samples 17901 unjoint_samples 17900 joint_samples 54 [266611, 1046346] processed_samples 17902 unjoint_samples 17900 joint_samples 53 [1047969, 903033] processed_samples 17902 unjoint_samples 17900 joint_samples 54 [127258, 1046568] processed_samples 17902 unjoint_samples 17900 joint_samples 52 [672743, 1046900] processed_samples 17901 unjoint_samples 17900 joint_samples 54 [936743, 1042309] [h264 @ 0x561ea4492980] mmco: unref short failure processed_samples 17902 unjoint_samples 17900 joint_samples 53 [1047969, 903033] [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure processed_samples 18000 unjoint_samples 18000 joint_samples 54 [184916, 1035067] processed_samples 18000 unjoint_samples 18000 joint_samples 54 [184916, 1035067] processed_samples 18001 unjoint_samples 18000 joint_samples 54 [781455, 1039875] processed_samples 18001 unjoint_samples 18000 joint_samples 54 [781455, 1039875] processed_samples 18001 unjoint_samples 18000 joint_samples 54 [168129, 1013275] processed_samples 18001 unjoint_samples 18000 joint_samples 54 [168129, 1013275] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure processed_samples 18001 unjoint_samples 18000 joint_samples 55 [174846, 1047908] processed_samples 18001 unjoint_samples 18000 joint_samples 55 [174846, 1047908] processed_samples 18001 unjoint_samples 18000 joint_samples 54 [584971, 1046346] processed_samples 18002 unjoint_samples 18000 joint_samples 52 [964530, 1046900] processed_samples 18001 unjoint_samples 18000 joint_samples 54 [584971, 1046346] processed_samples 18002 unjoint_samples 18000 joint_samples 52 [964530, 1046900] [h264 @ 0x55e6d1720f80] mmco: unref short failure processed_samples 18002 unjoint_samples 18000 joint_samples 54 [477254, 1046568] processed_samples 18002 unjoint_samples 18000 joint_samples 54 [262377, 1037102] processed_samples 18002 unjoint_samples 18000 joint_samples 54 [262377, 1037102] [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure processed_samples 18002 unjoint_samples 18000 joint_samples 54 [477254, 1046568] [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure processed_samples 18100 unjoint_samples 18100 joint_samples 54 [454708, 1035067] processed_samples 18101 unjoint_samples 18100 joint_samples 55 [110970, 1046508] processed_samples 18100 unjoint_samples 18100 joint_samples 54 [454708, 1035067] processed_samples 18101 unjoint_samples 18100 joint_samples 54 [443900, 1013275] processed_samples 18101 unjoint_samples 18100 joint_samples 55 [509440, 1047908] processed_samples 18101 unjoint_samples 18100 joint_samples 55 [110970, 1046508] processed_samples 18102 unjoint_samples 18100 joint_samples 53 [1027997, 210828] [h264 @ 0x55e6d1810ac0] mmco: unref short failure processed_samples 18101 unjoint_samples 18100 joint_samples 55 [509440, 1047908] processed_samples 18101 unjoint_samples 18100 joint_samples 54 [443900, 1013275] processed_samples 18102 unjoint_samples 18100 joint_samples 54 [564074, 1037102] processed_samples 18102 unjoint_samples 18100 joint_samples 53 [1027997, 210828] [h264 @ 0x561ea591cd00] mmco: unref short failure processed_samples 18102 unjoint_samples 18100 joint_samples 54 [564074, 1037102] processed_samples 18102 unjoint_samples 18100 joint_samples 54 [780530, 1046568] processed_samples 18101 unjoint_samples 18100 joint_samples 54 [856825, 1046346] processed_samples 18101 unjoint_samples 18100 joint_samples 54 [856825, 1046346] processed_samples 18102 unjoint_samples 18100 joint_samples 54 [780530, 1046568] [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea94d7900] stream 1, offset 0x14007e3: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6cd7f7a80] stream 1, offset 0x14007e3: partial file [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 18200 unjoint_samples 18200 joint_samples 54 [759433, 1035067] processed_samples 18200 unjoint_samples 18200 joint_samples 54 [759433, 1035067] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [551986, 1046508] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [551986, 1046508] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [782108, 1047908] processed_samples 18202 unjoint_samples 18200 joint_samples 55 [1042012, 74861] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [73656, 1046346] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [73656, 1046346] processed_samples 18201 unjoint_samples 18200 joint_samples 55 [782108, 1047908] processed_samples 18202 unjoint_samples 18200 joint_samples 55 [1042012, 74861] processed_samples 18202 unjoint_samples 18200 joint_samples 53 [1027997, 524387] processed_samples 18202 unjoint_samples 18200 joint_samples 53 [1027997, 524387] processed_samples 18202 unjoint_samples 18200 joint_samples 54 [837568, 1037102] processed_samples 18202 unjoint_samples 18200 joint_samples 54 [837568, 1037102] [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure processed_samples 18202 unjoint_samples 18200 joint_samples 54 [732610, 1013275] processed_samples 18202 unjoint_samples 18200 joint_samples 54 [732610, 1013275] [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x561ea142a9c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure processed_samples 18300 unjoint_samples 18300 joint_samples 55 [114136, 1047034] processed_samples 18302 unjoint_samples 18300 joint_samples 53 [1032588, 1037851] processed_samples 18301 unjoint_samples 18300 joint_samples 56 [1026659, 211365] processed_samples 18301 unjoint_samples 18300 joint_samples 55 [471891, 1046346] processed_samples 18300 unjoint_samples 18300 joint_samples 55 [114136, 1047034] [h264 @ 0x55e6d4750cc0] mmco: unref short failure processed_samples 18301 unjoint_samples 18300 joint_samples 56 [1026659, 211365] processed_samples 18302 unjoint_samples 18300 joint_samples 55 [1042012, 427515] processed_samples 18302 unjoint_samples 18300 joint_samples 53 [1032588, 1037851] processed_samples 18301 unjoint_samples 18300 joint_samples 55 [471891, 1046346] processed_samples 18302 unjoint_samples 18300 joint_samples 55 [1046681, 111732] processed_samples 18302 unjoint_samples 18300 joint_samples 54 [997672, 1013275] processed_samples 18301 unjoint_samples 18300 joint_samples 55 [824917, 1046508] [h264 @ 0x561ea3a4af00] mmco: unref short failure processed_samples 18302 unjoint_samples 18300 joint_samples 55 [1042012, 427515] processed_samples 18301 unjoint_samples 18300 joint_samples 55 [824917, 1046508] processed_samples 18302 unjoint_samples 18300 joint_samples 55 [1046681, 111732] processed_samples 18302 unjoint_samples 18300 joint_samples 54 [997672, 1013275] [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure processed_samples 18400 unjoint_samples 18400 joint_samples 55 [379762, 1047034] processed_samples 18400 unjoint_samples 18400 joint_samples 55 [379762, 1047034] processed_samples 18401 unjoint_samples 18400 joint_samples 56 [102646, 1046508] processed_samples 18401 unjoint_samples 18400 joint_samples 56 [102646, 1046508] processed_samples 18401 unjoint_samples 18400 joint_samples 56 [1026659, 617973] processed_samples 18401 unjoint_samples 18400 joint_samples 56 [1026659, 617973] processed_samples 18402 unjoint_samples 18400 joint_samples 55 [1042012, 842816] processed_samples 18402 unjoint_samples 18400 joint_samples 55 [1042012, 842816] [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure processed_samples 18402 unjoint_samples 18400 joint_samples 54 [314101, 1040157] processed_samples 18402 unjoint_samples 18400 joint_samples 54 [314101, 1040157] processed_samples 18402 unjoint_samples 18400 joint_samples 55 [1046681, 514907] processed_samples 18402 unjoint_samples 18400 joint_samples 55 [1046681, 514907] processed_samples 18402 unjoint_samples 18400 joint_samples 55 [494468, 1045459] processed_samples 18402 unjoint_samples 18400 joint_samples 55 [494468, 1045459] processed_samples 18401 unjoint_samples 18400 joint_samples 55 [839220, 1046346] processed_samples 18401 unjoint_samples 18400 joint_samples 55 [839220, 1046346] [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4521040] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6ce343140] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea1b43f40] mmco: unref short failure [h264 @ 0x55e6cdcc9f80] mmco: unref short failure [h264 @ 0x55e6cdcc9f80] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x55e6d15614c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure processed_samples 18500 unjoint_samples 18500 joint_samples 55 [733829, 1047034] [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure processed_samples 18500 unjoint_samples 18500 joint_samples 55 [733829, 1047034] processed_samples 18501 unjoint_samples 18500 joint_samples 56 [404680, 1046508] processed_samples 18501 unjoint_samples 18500 joint_samples 56 [404680, 1046508] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure processed_samples 18501 unjoint_samples 18500 joint_samples 56 [1046278, 68202] processed_samples 18501 unjoint_samples 18500 joint_samples 56 [1046278, 68202] processed_samples 18502 unjoint_samples 18500 joint_samples 54 [691067, 1040157] processed_samples 18502 unjoint_samples 18500 joint_samples 54 [691067, 1040157] processed_samples 18502 unjoint_samples 18500 joint_samples 56 [1042012, 96482] processed_samples 18502 unjoint_samples 18500 joint_samples 56 [1042012, 96482] processed_samples 18501 unjoint_samples 18500 joint_samples 56 [1026659, 928594] processed_samples 18501 unjoint_samples 18500 joint_samples 56 [1026659, 928594] processed_samples 18502 unjoint_samples 18500 joint_samples 55 [1046681, 756675] processed_samples 18502 unjoint_samples 18500 joint_samples 55 [1046681, 756675] processed_samples 18502 unjoint_samples 18500 joint_samples 55 [829792, 1045459] processed_samples 18502 unjoint_samples 18500 joint_samples 55 [829792, 1045459] [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea4a6d940] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure processed_samples 18600 unjoint_samples 18600 joint_samples 56 [1036320, 36100] processed_samples 18600 unjoint_samples 18600 joint_samples 56 [1036320, 36100] processed_samples 18601 unjoint_samples 18600 joint_samples 56 [1046278, 416982] processed_samples 18601 unjoint_samples 18600 joint_samples 56 [1046278, 416982] processed_samples 18602 unjoint_samples 18600 joint_samples 56 [1017296, 87009] processed_samples 18602 unjoint_samples 18600 joint_samples 56 [1017296, 87009] processed_samples 18601 unjoint_samples 18600 joint_samples 57 [1045498, 111568] processed_samples 18601 unjoint_samples 18600 joint_samples 57 [1045498, 111568] processed_samples 18601 unjoint_samples 18600 joint_samples 56 [668163, 1046508] processed_samples 18601 unjoint_samples 18600 joint_samples 56 [668163, 1046508] processed_samples 18602 unjoint_samples 18600 joint_samples 56 [13554, 1046527] processed_samples 18602 unjoint_samples 18600 joint_samples 56 [13554, 1046527] processed_samples 18602 unjoint_samples 18600 joint_samples 54 [987688, 1040157] processed_samples 18602 unjoint_samples 18600 joint_samples 54 [987688, 1040157] processed_samples 18602 unjoint_samples 18600 joint_samples 56 [1042012, 414081] processed_samples 18602 unjoint_samples 18600 joint_samples 56 [1042012, 414081] [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1036320, 292617] processed_samples 18701 unjoint_samples 18700 joint_samples 56 [932320, 1046508] processed_samples 18702 unjoint_samples 18700 joint_samples 55 [198196, 1044497] processed_samples 18701 unjoint_samples 18700 joint_samples 57 [1045498, 518352] processed_samples 18702 unjoint_samples 18700 joint_samples 56 [1017296, 392615] processed_samples 18702 unjoint_samples 18700 joint_samples 56 [1042012, 743118] processed_samples 18701 unjoint_samples 18700 joint_samples 56 [1046278, 814138] processed_samples 18702 unjoint_samples 18700 joint_samples 56 [351065, 1046527] processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1036320, 292617] [h264 @ 0x561ea238d980] mmco: unref short failure processed_samples 18702 unjoint_samples 18700 joint_samples 55 [198196, 1044497] processed_samples 18701 unjoint_samples 18700 joint_samples 57 [1045498, 518352] processed_samples 18701 unjoint_samples 18700 joint_samples 56 [1046278, 814138] processed_samples 18701 unjoint_samples 18700 joint_samples 56 [932320, 1046508] processed_samples 18702 unjoint_samples 18700 joint_samples 56 [1017296, 392615] processed_samples 18702 unjoint_samples 18700 joint_samples 56 [1042012, 743118] processed_samples 18702 unjoint_samples 18700 joint_samples 56 [351065, 1046527] [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x561ea923b600] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6d463d3c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure processed_samples 18800 unjoint_samples 18800 joint_samples 56 [1036320, 689638] processed_samples 18800 unjoint_samples 18800 joint_samples 56 [1036320, 689638] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [171195, 1046508] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [171195, 1046508] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [1047162, 83501] processed_samples 18802 unjoint_samples 18800 joint_samples 57 [1047499, 18744] processed_samples 18802 unjoint_samples 18800 joint_samples 57 [1047499, 18744] processed_samples 18802 unjoint_samples 18800 joint_samples 55 [453033, 1044497] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [1047162, 83501] processed_samples 18802 unjoint_samples 18800 joint_samples 55 [453033, 1044497] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [1045498, 863218] processed_samples 18802 unjoint_samples 18800 joint_samples 56 [1017296, 643171] processed_samples 18802 unjoint_samples 18800 joint_samples 56 [1017296, 643171] processed_samples 18802 unjoint_samples 18800 joint_samples 56 [652322, 1046527] processed_samples 18801 unjoint_samples 18800 joint_samples 57 [1045498, 863218] processed_samples 18802 unjoint_samples 18800 joint_samples 56 [652322, 1046527] [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x561ea4686140] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d5f8cdc0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure processed_samples 18901 unjoint_samples 18900 joint_samples 57 [488004, 1046508] [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure processed_samples 18901 unjoint_samples 18900 joint_samples 58 [136644, 1029487] processed_samples 18900 unjoint_samples 18900 joint_samples 56 [1036320, 979365] processed_samples 18901 unjoint_samples 18900 joint_samples 57 [1047162, 543750] processed_samples 18902 unjoint_samples 18900 joint_samples 57 [1047499, 379928] processed_samples 18902 unjoint_samples 18900 joint_samples 56 [1023930, 1046527] processed_samples 18902 unjoint_samples 18900 joint_samples 56 [1017296, 996235] processed_samples 18902 unjoint_samples 18900 joint_samples 55 [735267, 1044497] [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure processed_samples 18901 unjoint_samples 18900 joint_samples 57 [488004, 1046508] processed_samples 18900 unjoint_samples 18900 joint_samples 56 [1036320, 979365] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 18901 unjoint_samples 18900 joint_samples 58 [136644, 1029487] [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure processed_samples 18901 unjoint_samples 18900 joint_samples 57 [1047162, 543750] processed_samples 18902 unjoint_samples 18900 joint_samples 57 [1047499, 379928] processed_samples 18902 unjoint_samples 18900 joint_samples 56 [1023930, 1046527] processed_samples 18902 unjoint_samples 18900 joint_samples 55 [735267, 1044497] [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 18902 unjoint_samples 18900 joint_samples 56 [1017296, 996235] [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d18ad380] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure processed_samples 19001 unjoint_samples 19000 joint_samples 57 [736148, 1046508] processed_samples 19000 unjoint_samples 19000 joint_samples 57 [1040739, 351131] processed_samples 19000 unjoint_samples 19000 joint_samples 57 [1040739, 351131] processed_samples 19001 unjoint_samples 19000 joint_samples 57 [736148, 1046508] processed_samples 19001 unjoint_samples 19000 joint_samples 58 [374176, 1029487] processed_samples 19001 unjoint_samples 19000 joint_samples 58 [374176, 1029487] processed_samples 19002 unjoint_samples 19000 joint_samples 56 [1043543, 108480] processed_samples 19002 unjoint_samples 19000 joint_samples 56 [1043543, 108480] processed_samples 19002 unjoint_samples 19000 joint_samples 57 [1047141, 299843] processed_samples 19002 unjoint_samples 19000 joint_samples 57 [1047141, 299843] processed_samples 19002 unjoint_samples 19000 joint_samples 57 [1047499, 715763] processed_samples 19002 unjoint_samples 19000 joint_samples 57 [1033811, 310301] processed_samples 19002 unjoint_samples 19000 joint_samples 57 [1033811, 310301] processed_samples 19002 unjoint_samples 19000 joint_samples 57 [1047499, 715763] processed_samples 19001 unjoint_samples 19000 joint_samples 57 [1047162, 852402] processed_samples 19001 unjoint_samples 19000 joint_samples 57 [1047162, 852402] [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cdde8300] mmco: unref short failure [h264 @ 0x55e6cdde8300] mmco: unref short failure processed_samples 19100 unjoint_samples 19100 joint_samples 57 [1040739, 575743] processed_samples 19100 unjoint_samples 19100 joint_samples 57 [1040739, 575743] processed_samples 19101 unjoint_samples 19100 joint_samples 58 [782696, 1029487] processed_samples 19102 unjoint_samples 19100 joint_samples 57 [1047499, 1001486] processed_samples 19101 unjoint_samples 19100 joint_samples 58 [145200, 1044533] processed_samples 19102 unjoint_samples 19100 joint_samples 56 [1043543, 478941] processed_samples 19101 unjoint_samples 19100 joint_samples 57 [1014723, 1046508] processed_samples 19102 unjoint_samples 19100 joint_samples 57 [1047141, 616483] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 19101 unjoint_samples 19100 joint_samples 58 [782696, 1029487] processed_samples 19101 unjoint_samples 19100 joint_samples 58 [145200, 1044533] processed_samples 19102 unjoint_samples 19100 joint_samples 57 [1033811, 572453] processed_samples 19101 unjoint_samples 19100 joint_samples 57 [1014723, 1046508] processed_samples 19102 unjoint_samples 19100 joint_samples 56 [1043543, 478941] processed_samples 19102 unjoint_samples 19100 joint_samples 57 [1047499, 1001486] processed_samples 19102 unjoint_samples 19100 joint_samples 57 [1047141, 616483] [h264 @ 0x55e6cdfdfe40] mmco: unref short failure [h264 @ 0x55e6cdfdfe40] mmco: unref short failure processed_samples 19102 unjoint_samples 19100 joint_samples 57 [1033811, 572453] [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x561ea5922300] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x55e6cee01580] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure processed_samples 19201 unjoint_samples 19200 joint_samples 59 [1046395, 12727] processed_samples 19201 unjoint_samples 19200 joint_samples 59 [1046395, 12727] processed_samples 19201 unjoint_samples 19200 joint_samples 58 [247925, 1046508] processed_samples 19201 unjoint_samples 19200 joint_samples 58 [247925, 1046508] [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure processed_samples 19201 unjoint_samples 19200 joint_samples 58 [430047, 1044533] processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1040739, 824084] processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1040739, 824084] processed_samples 19201 unjoint_samples 19200 joint_samples 58 [430047, 1044533] processed_samples 19202 unjoint_samples 19200 joint_samples 58 [287258, 1047189] processed_samples 19202 unjoint_samples 19200 joint_samples 58 [287258, 1047189] processed_samples 19202 unjoint_samples 19200 joint_samples 57 [1047141, 922686] processed_samples 19202 unjoint_samples 19200 joint_samples 57 [1047141, 922686] processed_samples 19202 unjoint_samples 19200 joint_samples 56 [1043543, 796183] processed_samples 19202 unjoint_samples 19200 joint_samples 56 [1043543, 796183] processed_samples 19202 unjoint_samples 19200 joint_samples 57 [1033811, 896115] processed_samples 19202 unjoint_samples 19200 joint_samples 57 [1033811, 896115] [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6d42cbdc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea4bdeec0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure processed_samples 19300 unjoint_samples 19300 joint_samples 58 [1040739, 193914] processed_samples 19300 unjoint_samples 19300 joint_samples 58 [1040739, 193914] processed_samples 19301 unjoint_samples 19300 joint_samples 59 [1046395, 271014] processed_samples 19301 unjoint_samples 19300 joint_samples 59 [1046395, 271014] processed_samples 19301 unjoint_samples 19300 joint_samples 58 [475246, 1046508] processed_samples 19301 unjoint_samples 19300 joint_samples 58 [475246, 1046508] [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure processed_samples 19301 unjoint_samples 19300 joint_samples 58 [700396, 1044533] processed_samples 19302 unjoint_samples 19300 joint_samples 58 [1047312, 218473] processed_samples 19302 unjoint_samples 19300 joint_samples 58 [1047312, 218473] processed_samples 19301 unjoint_samples 19300 joint_samples 58 [700396, 1044533] processed_samples 19302 unjoint_samples 19300 joint_samples 58 [356666, 989930] processed_samples 19302 unjoint_samples 19300 joint_samples 58 [356666, 989930] processed_samples 19302 unjoint_samples 19300 joint_samples 58 [610170, 1047189] processed_samples 19302 unjoint_samples 19300 joint_samples 58 [610170, 1047189] [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure processed_samples 19302 unjoint_samples 19300 joint_samples 56 [1043543, 1022255] processed_samples 19302 unjoint_samples 19300 joint_samples 56 [1043543, 1022255] [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x55e6d1720f80] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d2e082c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x561ea91aaa00] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cf0e82c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure processed_samples 19400 unjoint_samples 19400 joint_samples 58 [1040739, 572496] processed_samples 19400 unjoint_samples 19400 joint_samples 58 [1040739, 572496] processed_samples 19402 unjoint_samples 19400 joint_samples 57 [327430, 1028977] processed_samples 19402 unjoint_samples 19400 joint_samples 57 [327430, 1028977] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [1046395, 541873] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [1046395, 541873] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [1044820, 69769] processed_samples 19401 unjoint_samples 19400 joint_samples 59 [1044820, 69769] processed_samples 19402 unjoint_samples 19400 joint_samples 58 [1047312, 490245] processed_samples 19402 unjoint_samples 19400 joint_samples 58 [658788, 989930] processed_samples 19402 unjoint_samples 19400 joint_samples 58 [658788, 989930] processed_samples 19402 unjoint_samples 19400 joint_samples 58 [1047312, 490245] processed_samples 19402 unjoint_samples 19400 joint_samples 58 [920996, 1047189] processed_samples 19402 unjoint_samples 19400 joint_samples 58 [920996, 1047189] processed_samples 19401 unjoint_samples 19400 joint_samples 58 [710589, 1046508] processed_samples 19401 unjoint_samples 19400 joint_samples 58 [710589, 1046508] [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure processed_samples 19500 unjoint_samples 19500 joint_samples 58 [1040739, 934234] [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure processed_samples 19502 unjoint_samples 19500 joint_samples 57 [641512, 1028977] processed_samples 19501 unjoint_samples 19500 joint_samples 59 [1044820, 418711] [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure processed_samples 19502 unjoint_samples 19500 joint_samples 59 [1046473, 130959] processed_samples 19502 unjoint_samples 19500 joint_samples 58 [1045277, 1027540] processed_samples 19501 unjoint_samples 19500 joint_samples 59 [1046395, 934132] processed_samples 19501 unjoint_samples 19500 joint_samples 58 [1005467, 1046508] processed_samples 19502 unjoint_samples 19500 joint_samples 58 [1047312, 715714] processed_samples 19500 unjoint_samples 19500 joint_samples 58 [1040739, 934234] [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure processed_samples 19502 unjoint_samples 19500 joint_samples 59 [1046473, 130959] processed_samples 19502 unjoint_samples 19500 joint_samples 57 [641512, 1028977] processed_samples 19501 unjoint_samples 19500 joint_samples 59 [1044820, 418711] processed_samples 19502 unjoint_samples 19500 joint_samples 58 [1045277, 1027540] [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure processed_samples 19501 unjoint_samples 19500 joint_samples 58 [1005467, 1046508] processed_samples 19502 unjoint_samples 19500 joint_samples 58 [1047312, 715714] processed_samples 19501 unjoint_samples 19500 joint_samples 59 [1046395, 934132] [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [h264 @ 0x561ea4ad0380] mmco: unref short failure [mov,mp4,m4a,3gp,3g2,mj2 @ 0x55e6cf3a7980] stream 0, offset 0x900676: partial file [mov,mp4,m4a,3gp,3g2,mj2 @ 0x561ea0e4c280] stream 0, offset 0x900676: partial file [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure processed_samples 19600 unjoint_samples 19600 joint_samples 59 [1047493, 299234] [h264 @ 0x55e6cf811340] mmco: unref short failure processed_samples 19600 unjoint_samples 19600 joint_samples 59 [1047493, 299234] processed_samples 19602 unjoint_samples 19600 joint_samples 59 [1045643, 377116] [h264 @ 0x561ea429e380] mmco: unref short failure processed_samples 19601 unjoint_samples 19600 joint_samples 60 [168639, 1046908] processed_samples 19602 unjoint_samples 19600 joint_samples 59 [1037020, 563308] processed_samples 19601 unjoint_samples 19600 joint_samples 59 [1044820, 742421] processed_samples 19602 unjoint_samples 19600 joint_samples 59 [1046473, 426743] [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure processed_samples 19602 unjoint_samples 19600 joint_samples 59 [1045643, 377116] processed_samples 19601 unjoint_samples 19600 joint_samples 60 [168639, 1046908] processed_samples 19602 unjoint_samples 19600 joint_samples 59 [1037020, 563308] processed_samples 19602 unjoint_samples 19600 joint_samples 57 [1033557, 1033272] processed_samples 19602 unjoint_samples 19600 joint_samples 59 [1046473, 426743] processed_samples 19601 unjoint_samples 19600 joint_samples 59 [1044820, 742421] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure processed_samples 19602 unjoint_samples 19600 joint_samples 57 [1033557, 1033272] processed_samples 19602 unjoint_samples 19600 joint_samples 58 [1047312, 948221] [h264 @ 0x55e6ce139800] mmco: unref short failure processed_samples 19602 unjoint_samples 19600 joint_samples 58 [1047312, 948221] [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea0f7dbc0] mmco: unref short failure [h264 @ 0x561ea0f7dbc0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x55e6cede0240] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x561ea1e4b940] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1047493, 592413] [h264 @ 0x55e6cd95f6c0] mmco: unref short failure processed_samples 19702 unjoint_samples 19700 joint_samples 59 [1037020, 922203] processed_samples 19702 unjoint_samples 19700 joint_samples 58 [1038292, 284168] processed_samples 19701 unjoint_samples 19700 joint_samples 60 [516012, 1046908] processed_samples 19702 unjoint_samples 19700 joint_samples 59 [235995, 1021819] processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1047493, 592413] processed_samples 19702 unjoint_samples 19700 joint_samples 59 [1045643, 643190] processed_samples 19701 unjoint_samples 19700 joint_samples 59 [1044820, 994944] [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1f86e80] mmco: unref short failure processed_samples 19702 unjoint_samples 19700 joint_samples 59 [1046473, 736051] processed_samples 19702 unjoint_samples 19700 joint_samples 59 [1037020, 922203] processed_samples 19702 unjoint_samples 19700 joint_samples 59 [235995, 1021819] processed_samples 19702 unjoint_samples 19700 joint_samples 58 [1038292, 284168] processed_samples 19701 unjoint_samples 19700 joint_samples 60 [516012, 1046908] processed_samples 19702 unjoint_samples 19700 joint_samples 59 [1045643, 643190] processed_samples 19702 unjoint_samples 19700 joint_samples 59 [1046473, 736051] processed_samples 19701 unjoint_samples 19700 joint_samples 59 [1044820, 994944] [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x55e6ce139800] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6d164f340] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1047493, 864123] [h264 @ 0x561ea17a22c0] mmco: unref short failure processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1047493, 864123] processed_samples 19802 unjoint_samples 19800 joint_samples 60 [181389, 976906] processed_samples 19801 unjoint_samples 19800 joint_samples 60 [1047103, 208695] processed_samples 19802 unjoint_samples 19800 joint_samples 58 [1038292, 618375] processed_samples 19802 unjoint_samples 19800 joint_samples 60 [1037020, 301549] processed_samples 19802 unjoint_samples 19800 joint_samples 59 [1046473, 1002538] [h264 @ 0x55e6ce622380] mmco: unref short failure processed_samples 19802 unjoint_samples 19800 joint_samples 59 [587015, 1021819] processed_samples 19801 unjoint_samples 19800 joint_samples 60 [886263, 1046908] processed_samples 19802 unjoint_samples 19800 joint_samples 60 [181389, 976906] processed_samples 19802 unjoint_samples 19800 joint_samples 60 [1037020, 301549] processed_samples 19801 unjoint_samples 19800 joint_samples 60 [1047103, 208695] processed_samples 19802 unjoint_samples 19800 joint_samples 58 [1038292, 618375] processed_samples 19801 unjoint_samples 19800 joint_samples 60 [886263, 1046908] processed_samples 19802 unjoint_samples 19800 joint_samples 59 [587015, 1021819] processed_samples 19802 unjoint_samples 19800 joint_samples 59 [1046473, 1002538] [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x55e6cd989800] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x561ea45d3300] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure processed_samples 19900 unjoint_samples 19900 joint_samples 60 [279421, 1046054] processed_samples 19901 unjoint_samples 19900 joint_samples 60 [1047103, 479436] processed_samples 19901 unjoint_samples 19900 joint_samples 61 [1003895, 262709] processed_samples 19901 unjoint_samples 19900 joint_samples 60 [1047103, 479436] processed_samples 19900 unjoint_samples 19900 joint_samples 60 [279421, 1046054] processed_samples 19901 unjoint_samples 19900 joint_samples 61 [1003895, 262709] processed_samples 19902 unjoint_samples 19900 joint_samples 60 [451205, 1046955] processed_samples 19902 unjoint_samples 19900 joint_samples 60 [522174, 976906] processed_samples 19902 unjoint_samples 19900 joint_samples 60 [451205, 1046955] processed_samples 19902 unjoint_samples 19900 joint_samples 60 [522174, 976906] processed_samples 19902 unjoint_samples 19900 joint_samples 59 [879594, 1021819] processed_samples 19902 unjoint_samples 19900 joint_samples 60 [1037020, 656509] processed_samples 19902 unjoint_samples 19900 joint_samples 59 [879594, 1021819] processed_samples 19902 unjoint_samples 19900 joint_samples 60 [1037020, 656509] processed_samples 19902 unjoint_samples 19900 joint_samples 58 [1038292, 827289] processed_samples 19902 unjoint_samples 19900 joint_samples 58 [1038292, 827289] [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d5cc97c0] mmco: unref short failure [h264 @ 0x561ea46fe9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x561ea42ec840] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea110ee40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure processed_samples 20000 unjoint_samples 20000 joint_samples 60 [655753, 1046054] processed_samples 20000 unjoint_samples 20000 joint_samples 60 [655753, 1046054] processed_samples 20002 unjoint_samples 20000 joint_samples 59 [125504, 1035232] processed_samples 20002 unjoint_samples 20000 joint_samples 59 [125504, 1035232] processed_samples 20001 unjoint_samples 20000 joint_samples 61 [1003895, 554741] processed_samples 20001 unjoint_samples 20000 joint_samples 61 [1003895, 554741] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [1037020, 1021795] processed_samples 20001 unjoint_samples 20000 joint_samples 60 [1047103, 816596] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [1037020, 1021795] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [117783, 1045668] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [117783, 1045668] processed_samples 20001 unjoint_samples 20000 joint_samples 60 [1047103, 816596] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [731960, 1046955] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [731960, 1046955] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [858074, 976906] processed_samples 20002 unjoint_samples 20000 joint_samples 60 [858074, 976906] [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6d19faf80] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea15a4780] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x561ea0f19780] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure processed_samples 20100 unjoint_samples 20100 joint_samples 60 [1030632, 1046054] processed_samples 20100 unjoint_samples 20100 joint_samples 60 [1030632, 1046054] processed_samples 20102 unjoint_samples 20100 joint_samples 61 [304771, 1031469] processed_samples 20102 unjoint_samples 20100 joint_samples 61 [249545, 1045962] processed_samples 20102 unjoint_samples 20100 joint_samples 61 [304771, 1031469] processed_samples 20102 unjoint_samples 20100 joint_samples 61 [249545, 1045962] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [1047103, 63734] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [1047103, 63734] processed_samples 20102 unjoint_samples 20100 joint_samples 59 [484191, 1035232] processed_samples 20102 unjoint_samples 20100 joint_samples 59 [484191, 1035232] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [1003895, 924522] processed_samples 20101 unjoint_samples 20100 joint_samples 61 [1003895, 924522] processed_samples 20102 unjoint_samples 20100 joint_samples 60 [488930, 1045668] processed_samples 20102 unjoint_samples 20100 joint_samples 60 [488930, 1045668] processed_samples 20102 unjoint_samples 20100 joint_samples 60 [978451, 1046955] processed_samples 20102 unjoint_samples 20100 joint_samples 60 [978451, 1046955] [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6ab44eac0] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea36e0880] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x561ea591c700] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d19da180] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6ce98b8c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6d6193600] mmco: unref short failure processed_samples 20200 unjoint_samples 20200 joint_samples 61 [204171, 1046054] processed_samples 20200 unjoint_samples 20200 joint_samples 61 [204171, 1046054] processed_samples 20201 unjoint_samples 20200 joint_samples 62 [138977, 1038064] processed_samples 20201 unjoint_samples 20200 joint_samples 62 [138977, 1038064] processed_samples 20202 unjoint_samples 20200 joint_samples 60 [780193, 1045668] processed_samples 20202 unjoint_samples 20200 joint_samples 60 [780193, 1045668] processed_samples 20202 unjoint_samples 20200 joint_samples 61 [1018877, 296773] processed_samples 20202 unjoint_samples 20200 joint_samples 61 [1018877, 296773] processed_samples 20201 unjoint_samples 20200 joint_samples 61 [1047103, 526213] processed_samples 20201 unjoint_samples 20200 joint_samples 61 [1047103, 526213] processed_samples 20202 unjoint_samples 20200 joint_samples 61 [508865, 1045962] processed_samples 20202 unjoint_samples 20200 joint_samples 61 [508865, 1045962] processed_samples 20202 unjoint_samples 20200 joint_samples 61 [736308, 1031469] processed_samples 20202 unjoint_samples 20200 joint_samples 61 [736308, 1031469] processed_samples 20202 unjoint_samples 20200 joint_samples 59 [801639, 1035232] processed_samples 20202 unjoint_samples 20200 joint_samples 59 [801639, 1035232] [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1668f00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x55e6d1f0d4c0] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6cec12f80] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x55e6cffe1c80] mmco: unref short failure [h264 @ 0x561ea4251980] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6cfce0140] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea42eb580] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6ced74880] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x55e6ce98dc00] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure processed_samples 20302 unjoint_samples 20300 joint_samples 60 [1019239, 144609] processed_samples 20301 unjoint_samples 20300 joint_samples 62 [468946, 1038064] processed_samples 20300 unjoint_samples 20300 joint_samples 61 [491450, 1046054] processed_samples 20302 unjoint_samples 20300 joint_samples 61 [1018877, 548192] processed_samples 20302 unjoint_samples 20300 joint_samples 60 [1019239, 144609] processed_samples 20302 unjoint_samples 20300 joint_samples 61 [59675, 1047787] processed_samples 20300 unjoint_samples 20300 joint_samples 61 [491450, 1046054] processed_samples 20301 unjoint_samples 20300 joint_samples 61 [1047103, 847445] processed_samples 20301 unjoint_samples 20300 joint_samples 61 [1047103, 847445] processed_samples 20302 unjoint_samples 20300 joint_samples 61 [1018877, 548192] processed_samples 20302 unjoint_samples 20300 joint_samples 61 [59675, 1047787] processed_samples 20301 unjoint_samples 20300 joint_samples 62 [468946, 1038064] processed_samples 20302 unjoint_samples 20300 joint_samples 61 [893863, 1045962] processed_samples 20302 unjoint_samples 20300 joint_samples 61 [1034095, 1035922] processed_samples 20302 unjoint_samples 20300 joint_samples 61 [893863, 1045962] processed_samples 20302 unjoint_samples 20300 joint_samples 61 [1034095, 1035922] [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x55e6d5b9bf40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x561ea119a540] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea39e4840] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x55e6cf301c80] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x55e6d16cd300] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cf8a8940] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x561ea1ca0880] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x55e6d1810ac0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x561ea12d2dc0] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x55e6ce2cfe40] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea548af40] mmco: unref short failure [h264 @ 0x55e6cef61b40] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure processed_samples 20400 unjoint_samples 20400 joint_samples 61 [862076, 1046054] processed_samples 20402 unjoint_samples 20400 joint_samples 62 [1048168, 342758] processed_samples 20402 unjoint_samples 20400 joint_samples 60 [1019239, 536306] processed_samples 20400 unjoint_samples 20400 joint_samples 61 [862076, 1046054] processed_samples 20402 unjoint_samples 20400 joint_samples 60 [1019239, 536306] processed_samples 20402 unjoint_samples 20400 joint_samples 62 [1048168, 342758] [h264 @ 0x561ea1be5dc0] mmco: unref short failure processed_samples 20402 unjoint_samples 20400 joint_samples 62 [1047299, 181231] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [1047103, 256160] [h264 @ 0x55e6cdfdfe40] mmco: unref short failure processed_samples 20402 unjoint_samples 20400 joint_samples 62 [1047299, 181231] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [1047103, 256160] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [835179, 1038064] processed_samples 20401 unjoint_samples 20400 joint_samples 62 [835179, 1038064] processed_samples 20402 unjoint_samples 20400 joint_samples 61 [1018877, 892008] processed_samples 20402 unjoint_samples 20400 joint_samples 61 [1018877, 892008] processed_samples 20402 unjoint_samples 20400 joint_samples 61 [431313, 1047787] processed_samples 20402 unjoint_samples 20400 joint_samples 61 [431313, 1047787] [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea060ea80] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x561ea1be5dc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b608c0] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6cec0ee00] mmco: unref short failure [h264 @ 0x55e6d6719000] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x55e6cddc4180] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea4ca4640] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x55e6ce2c8d00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d0616a40] mmco: unref short failure [h264 @ 0x55e6cea92e40] mmco: unref short failure [h264 @ 0x561ea4b5b940] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4c60e40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure processed_samples 20500 unjoint_samples 20500 joint_samples 62 [69735, 1046054] processed_samples 20500 unjoint_samples 20500 joint_samples 62 [69735, 1046054] processed_samples 20501 unjoint_samples 20500 joint_samples 62 [1047103, 587150] processed_samples 20501 unjoint_samples 20500 joint_samples 63 [1027342, 120251] processed_samples 20501 unjoint_samples 20500 joint_samples 62 [1047103, 587150] processed_samples 20501 unjoint_samples 20500 joint_samples 63 [1027342, 120251] processed_samples 20502 unjoint_samples 20500 joint_samples 62 [135569, 1038496] processed_samples 20502 unjoint_samples 20500 joint_samples 62 [135569, 1038496] processed_samples 20502 unjoint_samples 20500 joint_samples 62 [1047299, 539034] processed_samples 20502 unjoint_samples 20500 joint_samples 62 [1047299, 539034] processed_samples 20502 unjoint_samples 20500 joint_samples 61 [715811, 1047787] processed_samples 20502 unjoint_samples 20500 joint_samples 61 [715811, 1047787] processed_samples 20502 unjoint_samples 20500 joint_samples 62 [1048168, 660308] processed_samples 20502 unjoint_samples 20500 joint_samples 62 [1048168, 660308] processed_samples 20502 unjoint_samples 20500 joint_samples 60 [1019239, 876256] processed_samples 20502 unjoint_samples 20500 joint_samples 60 [1019239, 876256] [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x561ea0b9db00] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x561ea469e3c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea4681800] mmco: unref short failure [h264 @ 0x55e6d199ac00] mmco: unref short failure [h264 @ 0x561ea27ac400] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x561ea12bff40] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea25b2580] mmco: unref short failure [h264 @ 0x561ea4813cc0] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6d16cf440] mmco: unref short failure [h264 @ 0x55e6cd948700] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x561ea27ab6c0] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x55e6d1a48d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6d28acdc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6ceabf8c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea238d980] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea0ce7cc0] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x561ea1049640] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cd95f6c0] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x55e6d180b300] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x55e6d1ec6800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0a5b3c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x561ea429e380] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d4750cc0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea591cd00] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x561ea4492980] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x561ea1b67a00] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x561ea4b52b40] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x55e6ce8d5100] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x561ea17a22c0] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure processed_samples 20602 unjoint_samples 20600 joint_samples 61 [1045027, 127791] processed_samples 20601 unjoint_samples 20600 joint_samples 63 [1027342, 389069] processed_samples 20600 unjoint_samples 20600 joint_samples 62 [493983, 1046054] processed_samples 20602 unjoint_samples 20600 joint_samples 61 [1045027, 127791] processed_samples 20600 unjoint_samples 20600 joint_samples 62 [493983, 1046054] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [1047299, 802293] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [1029424, 53213] processed_samples 20601 unjoint_samples 20600 joint_samples 62 [1047103, 989933] processed_samples 20601 unjoint_samples 20600 joint_samples 63 [1027342, 389069] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [405417, 1038496] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [1047299, 802293] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [1048168, 967805] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [1029424, 53213] processed_samples 20601 unjoint_samples 20600 joint_samples 62 [1047103, 989933] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [405417, 1038496] processed_samples 20602 unjoint_samples 20600 joint_samples 62 [1048168, 967805] [h264 @ 0x561ea359f500] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x561ea4bd3400] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ceb09e40] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea5c880c0] mmco: unref short failure [h264 @ 0x561ea47fa280] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdfe92c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x561ea0ad5140] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x561ea4ce19c0] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x55e6ce378200] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cd91c9c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x561ea433ab80] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea48b7d00] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x55e6cdbf6580] mmco: unref short failure [h264 @ 0x561ea434c4c0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6cf0af8c0] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x561ea30e4f40] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6d4292640] mmco: unref short failure [h264 @ 0x55e6ce443040] mmco: unref short failure [h264 @ 0x561ea4e60d40] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x55e6ce721500] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea27bbf00] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x561ea0f7dbc0] mmco: unref short failure [h264 @ 0x55e6cdce6a40] mmco: unref short failure processed_samples 20701 unjoint_samples 20700 joint_samples 63 [1047103, 201590] processed_samples 20702 unjoint_samples 20700 joint_samples 61 [1045027, 500733] processed_samples 20700 unjoint_samples 20700 joint_samples 62 [847874, 1046054] processed_samples 20701 unjoint_samples 20700 joint_samples 63 [1047103, 201590] processed_samples 20702 unjoint_samples 20700 joint_samples 61 [1045027, 500733] processed_samples 20702 unjoint_samples 20700 joint_samples 63 [1047299, 194005] processed_samples 20702 unjoint_samples 20700 joint_samples 63 [1047299, 194005] processed_samples 20700 unjoint_samples 20700 joint_samples 62 [847874, 1046054] processed_samples 20702 unjoint_samples 20700 joint_samples 62 [1029424, 413531] processed_samples 20702 unjoint_samples 20700 joint_samples 62 [1029424, 413531] processed_samples 20702 unjoint_samples 20700 joint_samples 63 [197870, 1041758] processed_samples 20702 unjoint_samples 20700 joint_samples 63 [197870, 1041758] processed_samples 20701 unjoint_samples 20700 joint_samples 63 [1027342, 639189] processed_samples 20701 unjoint_samples 20700 joint_samples 63 [1027342, 639189] processed_samples 20702 unjoint_samples 20700 joint_samples 62 [744615, 1038496] [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea64a27c0] mmco: unref short failure processed_samples 20702 unjoint_samples 20700 joint_samples 62 [744615, 1038496] [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6ce4a0e00] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x561ea3ffc480] mmco: unref short failure [h264 @ 0x55e6ce9550c0] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [h264 @ 0x561ea3a4af00] mmco: unref short failure [h264 @ 0x55e6cdb2f440] mmco: unref short failure [h264 @ 0x561ea4b57100] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cdc30880] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x55e6cd8924c0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x561ea4868bc0] mmco: unref short failure [h264 @ 0x55e6d28be940] mmco: unref short failure [h264 @ 0x561ea05da800] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x561ea0923a40] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x55e6d0971040] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x561ea0789140] mmco: unref short failure [h264 @ 0x55e6ce622380] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x55e6cec11700] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x561ea0d541c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x55e6cf5ec7c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea437d0c0] mmco: unref short failure [h264 @ 0x561ea080ecc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d18e3740] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x55e6cf811340] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x55e6d1fb24c0] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x561ea4c57c80] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x55e6d1725dc0] mmco: unref short failure [h264 @ 0x561ea23d1800] mmco: unref short failure [h264 @ 0x561ea47f8fc0] mmco: unref short failure [h264 @ 0x55e6cea6c680] mmco: unref short failure [h264 @ 0x55e6cdf59cc0] mmco: unref short failure [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2'[Errno 19] No such device: 'data_2'[Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2'[Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [Errno 19] No such device: 'data_2' [2024-12-01 22:43:54,174] torch.distributed.elastic.agent.server.api: [ERROR] Error waiting on exit barrier. Elapsed: 300.0093536376953 seconds + set +x