+ echo Logging output to /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//log_node7.txt
Logging output to /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//log_node7.txt
+ export ASCEND_PROCESS_LOG_PATH=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/7
+ ASCEND_PROCESS_LOG_PATH=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/7
+ mkdir -p /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//ascend/7
+ DATA_PATH=/local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml
+ TOKENIZER_PATH=/data_4/models/Qwen/Qwen2.5-14B-Instruct/
+ CKPT_LOAD_DIR=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/
+ VIT_CKPT_LOAD_DIR=/
+ CKPT_SAVE_DIR=/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743//
+ rsync -avh /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743/
sending incremental file list

sent 71 bytes  received 12 bytes  166.00 bytes/sec
total size is 23.84K  speedup is 287.17
+ cd /local_disk/cognitron_vl/
+ rm -fr datasets
+ mkdir -p datasets
+ ln -s /data/data/ datasets/CV
+ ln -s /data/data/LLM datasets/LLM
+ ln -s /data/data/LMM datasets/LMM
+ source /local_disk/cognitron_vl//scripts/set_env_mg_npu.sh
++ source /usr/local/Ascend/driver/bin/setenv.bash
+++ DEP_INFO_FILE=/etc/ascend_install.info
+++ [[ -f /etc/ascend_install.info ]]
+++ . /etc/ascend_install.info
+++ DRV_LIB64_COMMON_LDPATH=/driver/lib64/common
+++ DRV_LIB64_DRV_LDPATH=/driver/lib64/driver
+++ DRV_LIB64_LDPATH=/driver/lib64
+++ export LD_LIBRARY_PATH=/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial:
+++ LD_LIBRARY_PATH=/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial:
+++ export PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin
+++ PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin
++ source /usr/local/Ascend/ascend-toolkit/set_env.sh
+++ export LD_LIBRARY_PATH=/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial:
+++ LD_LIBRARY_PATH=/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial:
+++ export ASCEND_TOOLKIT_HOME=/usr/local/Ascend/ascend-toolkit/latest
+++ ASCEND_TOOLKIT_HOME=/usr/local/Ascend/ascend-toolkit/latest
++++ arch
+++ export LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial:
+++ LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial:
+++ export LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial:
+++ LD_LIBRARY_PATH=/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/driver/lib64/common:/driver/lib64/driver:/driver/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64:/usr/local/Ascend/ascend-toolkit/latest/tools/aml/lib64/plugin:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe/op_tiling/lib/linux/x86_64:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/lib/x86_64-linux-gnu/hdf5/serial:
+++ export PYTHONPATH=/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:
+++ PYTHONPATH=/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:
+++ export PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin
+++ PATH=/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/bin:/usr/local/Ascend/ascend-toolkit/latest/compiler/ccec_compiler/bin:/usr/local/Ascend/ascend-toolkit/latest/tools/ccec_compiler/bin:/root/miniconda3/envs/py38/bin:/root/miniconda3/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/sbin:/usr/local/bin
+++ export ASCEND_AICPU_PATH=/usr/local/Ascend/ascend-toolkit/latest
+++ ASCEND_AICPU_PATH=/usr/local/Ascend/ascend-toolkit/latest
+++ export ASCEND_OPP_PATH=/usr/local/Ascend/ascend-toolkit/latest/opp
+++ ASCEND_OPP_PATH=/usr/local/Ascend/ascend-toolkit/latest/opp
+++ export TOOLCHAIN_HOME=/usr/local/Ascend/ascend-toolkit/latest/toolkit
+++ TOOLCHAIN_HOME=/usr/local/Ascend/ascend-toolkit/latest/toolkit
+++ export ASCEND_HOME_PATH=/usr/local/Ascend/ascend-toolkit/latest
+++ ASCEND_HOME_PATH=/usr/local/Ascend/ascend-toolkit/latest
++ export HCCL_CONNECT_TIMEOUT=7200
++ HCCL_CONNECT_TIMEOUT=7200
++ export HCCL_EXEC_TIMEOUT=7200
++ HCCL_EXEC_TIMEOUT=7200
++ export COMBINED_ENABLE=1
++ COMBINED_ENABLE=1
++ export MULTI_STREAM_MEMORY_REUSE=1
++ MULTI_STREAM_MEMORY_REUSE=1
++ export HCCL_RDMA_TC=160
++ HCCL_RDMA_TC=160
++ export HCCL_RDMA_SL=5
++ HCCL_RDMA_SL=5
++ export HCCL_INTRA_PCIE_ENABLE=0
++ HCCL_INTRA_PCIE_ENABLE=0
++ export HCCL_INTRA_ROCE_ENABLE=1
++ HCCL_INTRA_ROCE_ENABLE=1
++ export HCCL_RDMA_TIMEOUT=20
++ HCCL_RDMA_TIMEOUT=20
++ export INF_NAN_MODE_ENABLE=1
++ INF_NAN_MODE_ENABLE=1
++ export DISTRIBUTED_BACKEND=hccl
++ DISTRIBUTED_BACKEND=hccl
++ export ASCEND_LAUNCH_BLOCKING=0
++ ASCEND_LAUNCH_BLOCKING=0
++ export ASCEND_SLOG_PRINT_TO_STDOUT=0
++ ASCEND_SLOG_PRINT_TO_STDOUT=0
++ export ASCEND_GLOBAL_LOG_LEVEL=3
++ ASCEND_GLOBAL_LOG_LEVEL=3
++ export ASCEND_GLOBAL_EVENT_ENABLE=0
++ ASCEND_GLOBAL_EVENT_ENABLE=0
++ export TASK_QUEUE_ENABLE=1
++ TASK_QUEUE_ENABLE=1
++ export PTCOPY_ENABLE=1
++ PTCOPY_ENABLE=1
++ export COMBINED_ENABLE=1
++ COMBINED_ENABLE=1
++ export DYNAMIC_OP=ADD#MUL
++ DYNAMIC_OP=ADD#MUL
++ export HCCL_WHITELIST_DISABLE=1
++ HCCL_WHITELIST_DISABLE=1
++ export HCCL_CONNECT_TIMEOUT=7200
++ HCCL_CONNECT_TIMEOUT=7200
++ export HCCL_WHITELIST_DISABLE=1
++ HCCL_WHITELIST_DISABLE=1
++ export CUDA_DEVICE_MAX_CONNECTIONS=1
++ CUDA_DEVICE_MAX_CONNECTIONS=1
++ pip3 install --no-index --find-links=/data/software/ -r requirements_npu.txt
Looking in links: /data/software/
Processing data/software/expecttest-0.2.1-py3-none-any.whl (from -r requirements_npu.txt (line 1))
Requirement already satisfied: peft in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 2)) (0.7.0)
Processing data/software/XlsxWriter-3.2.0-py3-none-any.whl (from -r requirements_npu.txt (line 3))
Requirement already satisfied: termcolor in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 4)) (2.4.0)
Requirement already satisfied: tabulate in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 5)) (0.9.0)
Processing data/software/tiktoken-0.7.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from -r requirements_npu.txt (line 6))
Requirement already satisfied: matplotlib in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 7)) (3.7.5)
Processing data/software/datasets-3.0.0-py3-none-any.whl (from -r requirements_npu.txt (line 8))
Requirement already satisfied: einops in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 9)) (0.7.0)
Processing data/software/pybind11-2.13.6-py3-none-any.whl (from -r requirements_npu.txt (line 10))
Requirement already satisfied: tensorboardX in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 11)) (2.6.2.2)
Processing data/software/pyarrow-17.0.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from -r requirements_npu.txt (line 12))
Requirement already satisfied: transformers>=4.40.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 13)) (4.40.1)
Requirement already satisfied: deepspeed>=0.14.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 14)) (0.14.5)
Processing data/software/accelerate-0.34.2-py3-none-any.whl (from -r requirements_npu.txt (line 15))
Requirement already satisfied: timm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from -r requirements_npu.txt (line 16)) (0.9.16)
Processing data/software/flask-3.0.3-py3-none-any.whl (from -r requirements_npu.txt (line 17))
Processing data/software/Flask_RESTful-0.3.10-py2.py3-none-any.whl (from -r requirements_npu.txt (line 18))
Processing data/software/decord-0.6.0-py3-none-manylinux2010_x86_64.whl (from -r requirements_npu.txt (line 19))
Processing data/software/natsort-8.4.0-py3-none-any.whl (from -r requirements_npu.txt (line 20))
Requirement already satisfied: numpy>=1.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (1.24.4)
Requirement already satisfied: packaging>=20.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (23.2)
Requirement already satisfied: psutil in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (5.9.8)
Requirement already satisfied: pyyaml in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (5.4.1)
Requirement already satisfied: torch>=1.13.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (2.1.0+cpu)
Requirement already satisfied: tqdm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (4.66.2)
Requirement already satisfied: safetensors in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (0.4.2)
Requirement already satisfied: huggingface-hub>=0.17.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft->-r requirements_npu.txt (line 2)) (0.20.3)
Requirement already satisfied: regex>=2022.1.18 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tiktoken->-r requirements_npu.txt (line 6)) (2023.12.25)
Requirement already satisfied: requests>=2.26.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tiktoken->-r requirements_npu.txt (line 6)) (2.31.0)
Requirement already satisfied: contourpy>=1.0.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (1.1.1)
Requirement already satisfied: cycler>=0.10 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (0.12.1)
Requirement already satisfied: fonttools>=4.22.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (4.49.0)
Requirement already satisfied: kiwisolver>=1.0.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (1.4.5)
Requirement already satisfied: pillow>=6.2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (10.2.0)
Requirement already satisfied: pyparsing>=2.3.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (3.1.1)
Requirement already satisfied: python-dateutil>=2.7 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (2.8.2)
Requirement already satisfied: importlib-resources>=3.2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from matplotlib->-r requirements_npu.txt (line 7)) (6.1.2)
Requirement already satisfied: filelock in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.13.1)
Requirement already satisfied: dill<0.3.9,>=0.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (0.3.7)
Requirement already satisfied: pandas in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2.0.3)
Processing data/software/requests-2.32.3-py3-none-any.whl (from tiktoken->-r requirements_npu.txt (line 6))
Processing data/software/tqdm-4.67.1-py3-none-any.whl (from peft->-r requirements_npu.txt (line 2))
Requirement already satisfied: xxhash in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.4.1)
Requirement already satisfied: multiprocess in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (0.70.15)
Requirement already satisfied: fsspec<=2024.6.1,>=2023.1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from fsspec[http]<=2024.6.1,>=2023.1.0->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2023.10.0)
Requirement already satisfied: aiohttp in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets>=2.21.0->-r requirements_npu.txt (line 8)) (3.9.3)
Processing data/software/huggingface_hub-0.26.2-py3-none-any.whl (from peft->-r requirements_npu.txt (line 2))
Requirement already satisfied: protobuf>=3.20 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from tensorboardX->-r requirements_npu.txt (line 11)) (4.25.3)
Requirement already satisfied: tokenizers<0.20,>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers>=4.40.1->-r requirements_npu.txt (line 13)) (0.19.1)
Requirement already satisfied: hjson in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (3.1.0)
Requirement already satisfied: ninja in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (1.11.1.1)
Requirement already satisfied: nvidia-ml-py in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (12.560.30)
Requirement already satisfied: py-cpuinfo in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (9.0.0)
Requirement already satisfied: pydantic in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from deepspeed>=0.14.2->-r requirements_npu.txt (line 14)) (1.10.15)
Processing data/software/safetensors-0.4.5-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from peft->-r requirements_npu.txt (line 2))
Requirement already satisfied: torchvision in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from timm->-r requirements_npu.txt (line 16)) (0.16.0)
Requirement already satisfied: Werkzeug>=3.0.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (3.0.1)
Requirement already satisfied: Jinja2>=3.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (3.1.3)
Processing data/software/itsdangerous-2.2.0-py3-none-any.whl (from flask->-r requirements_npu.txt (line 17))
Requirement already satisfied: click>=8.1.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (8.1.7)
Processing data/software/blinker-1.8.2-py3-none-any.whl (from flask->-r requirements_npu.txt (line 17))
Requirement already satisfied: importlib-metadata>=3.6.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask->-r requirements_npu.txt (line 17)) (7.0.1)
Processing data/software/aniso8601-9.0.1-py2.py3-none-any.whl (from flask_restful->-r requirements_npu.txt (line 18))
Requirement already satisfied: six>=1.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask_restful->-r requirements_npu.txt (line 18)) (1.16.0)
Requirement already satisfied: pytz in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from flask_restful->-r requirements_npu.txt (line 18)) (2024.1)
Requirement already satisfied: aiosignal>=1.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.3.1)
Requirement already satisfied: attrs>=17.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (23.2.0)
Requirement already satisfied: frozenlist>=1.1.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.4.1)
Requirement already satisfied: multidict<7.0,>=4.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (6.0.5)
Requirement already satisfied: yarl<2.0,>=1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (1.9.4)
Requirement already satisfied: async-timeout<5.0,>=4.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (4.0.3)
Requirement already satisfied: typing-extensions>=3.7.4.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from huggingface-hub>=0.17.0->peft->-r requirements_npu.txt (line 2)) (4.10.0)
Requirement already satisfied: zipp>=0.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from importlib-metadata>=3.6.0->flask->-r requirements_npu.txt (line 17)) (3.17.0)
Requirement already satisfied: MarkupSafe>=2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from Jinja2>=3.1.2->flask->-r requirements_npu.txt (line 17)) (2.1.5)
Requirement already satisfied: charset-normalizer<4,>=2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (3.3.2)
Requirement already satisfied: idna<4,>=2.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (3.6)
Requirement already satisfied: urllib3<3,>=1.21.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (1.26.18)
Requirement already satisfied: certifi>=2017.4.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests>=2.26.0->tiktoken->-r requirements_npu.txt (line 6)) (2024.2.2)
Requirement already satisfied: sympy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (1.4)
Requirement already satisfied: networkx in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (3.1)
Requirement already satisfied: tzdata>=2022.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets>=2.21.0->-r requirements_npu.txt (line 8)) (2024.1)
Requirement already satisfied: mpmath>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from sympy->torch>=1.13.0->peft->-r requirements_npu.txt (line 2)) (1.3.0)
DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063
Installing collected packages: aniso8601, xlsxwriter, tqdm, safetensors, requests, pybind11, pyarrow, natsort, itsdangerous, expecttest, decord, blinker, tiktoken, huggingface-hub, flask, flask_restful, accelerate, datasets
  Attempting uninstall: tqdm
    Found existing installation: tqdm 4.66.2
    Uninstalling tqdm-4.66.2:
      Successfully uninstalled tqdm-4.66.2
  Attempting uninstall: safetensors
    Found existing installation: safetensors 0.4.2
    Uninstalling safetensors-0.4.2:
      Successfully uninstalled safetensors-0.4.2
  Attempting uninstall: requests
    Found existing installation: requests 2.31.0
    Uninstalling requests-2.31.0:
      Successfully uninstalled requests-2.31.0
  Attempting uninstall: pyarrow
    Found existing installation: pyarrow 15.0.0
    Uninstalling pyarrow-15.0.0:
      Successfully uninstalled pyarrow-15.0.0
  Attempting uninstall: huggingface-hub
    Found existing installation: huggingface-hub 0.20.3
    Uninstalling huggingface-hub-0.20.3:
      Successfully uninstalled huggingface-hub-0.20.3
  Attempting uninstall: accelerate
    Found existing installation: accelerate 0.25.0
    Uninstalling accelerate-0.25.0:
      Successfully uninstalled accelerate-0.25.0
  Attempting uninstall: datasets
    Found existing installation: datasets 2.16.0
    Uninstalling datasets-2.16.0:
      Successfully uninstalled datasets-2.16.0
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
tikit 1.8.2.240926 requires dicttoxml==1.7.4, which is not installed.
tikit 1.8.2.240926 requires docopt==0.6.2, which is not installed.
tikit 1.8.2.240926 requires future==0.18.2, which is not installed.
tikit 1.8.2.240926 requires hdfs==2.6.0, which is not installed.
tikit 1.8.2.240926 requires pure-sasl==0.6.2, which is not installed.
tikit 1.8.2.240926 requires py4j==0.10.7, which is not installed.
tikit 1.8.2.240926 requires PyHive[hive]==0.6.4, which is not installed.
tikit 1.8.2.240926 requires pyjwt>=2.4.0, which is not installed.
tikit 1.8.2.240926 requires requests-kerberos>=0.14.0, which is not installed.
tikit 1.8.2.240926 requires sasl==0.3.1, which is not installed.
tikit 1.8.2.240926 requires thrift==0.15.0, which is not installed.
tikit 1.8.2.240926 requires thrift-sasl>=0.1.0, which is not installed.
tikit 1.8.2.240926 requires certifi==2021.10.8, but you have certifi 2024.2.2 which is incompatible.
tikit 1.8.2.240926 requires cos-python-sdk-v5==1.9.29, but you have cos-python-sdk-v5 1.9.26 which is incompatible.
tikit 1.8.2.240926 requires idna==3.3, but you have idna 3.6 which is incompatible.
tikit 1.8.2.240926 requires prettytable==2.5.0, but you have prettytable 3.11.0 which is incompatible.
tikit 1.8.2.240926 requires urllib3==1.26.7, but you have urllib3 1.26.18 which is incompatible.
tikit 1.8.2.240926 requires wcwidth==0.2.5, but you have wcwidth 0.2.13 which is incompatible.
Successfully installed accelerate-0.34.2 aniso8601-9.0.1 blinker-1.8.2 datasets-3.0.0 decord-0.6.0 expecttest-0.2.1 flask-3.0.3 flask_restful-0.3.10 huggingface-hub-0.26.2 itsdangerous-2.2.0 natsort-8.4.0 pyarrow-17.0.0 pybind11-2.13.6 requests-2.32.3 safetensors-0.4.5 tiktoken-0.7.0 tqdm-4.67.1 xlsxwriter-3.2.0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
++ return 0
+ MEGATRON_DIR=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0/
+ MINDSPEED_DIR=/local_disk/cognitron_vl//third_party/MindSpeed_core_r0.6.0/
+ MODELLINK_DIR=/local_disk/cognitron_vl//third_party/ModelLink/
+ pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0/
Looking in links: /data/software/
Obtaining file://local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0
  Installing build dependencies: started
  Installing build dependencies: finished with status 'done'
  Checking if build backend supports build_editable: started
  Checking if build backend supports build_editable: finished with status 'done'
  Getting requirements to build editable: started
  Getting requirements to build editable: finished with status 'done'
  Installing backend dependencies: started
  Installing backend dependencies: finished with status 'done'
  Preparing editable metadata (pyproject.toml): started
  Preparing editable metadata (pyproject.toml): finished with status 'done'
Building wheels for collected packages: megatron_core
  Building editable for megatron_core (pyproject.toml): started
  Building editable for megatron_core (pyproject.toml): finished with status 'done'
  Created wheel for megatron_core: filename=megatron_core-0.6.0-0.editable-cp38-cp38-linux_x86_64.whl size=8791 sha256=45264bd466d16535ad2060a555b895f6f1ef1411bdb5aac2bc508966f4578e95
  Stored in directory: /tmp/pip-ephem-wheel-cache-qz96g3li/wheels/54/9c/d1/d2015aa0c34e791e64d65d19395e5a9a5528f0c63fd519b9ff
Successfully built megatron_core
DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063
Installing collected packages: megatron_core
Successfully installed megatron_core-0.6.0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
+ pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/MindSpeed_core_r0.6.0/
Looking in links: /data/software/
Obtaining file://local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0
  Preparing metadata (setup.py): started
  Preparing metadata (setup.py): finished with status 'done'
WARNING: Error parsing requirements for tokenizers: [Errno 2] No such file or directory: '/root/miniconda3/envs/py38/lib/python3.8/site-packages/tokenizers-0.19.1.dist-info/METADATA'
WARNING: Error parsing requirements for transformers: [Errno 2] No such file or directory: '/root/miniconda3/envs/py38/lib/python3.8/site-packages/transformers-4.40.1.dist-info/METADATA'
DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063
Installing collected packages: mindspeed
  Running setup.py develop for mindspeed
Successfully installed mindspeed-0.6.0
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
+ pip3 install --no-index --find-links=/data/software/ -e /local_disk/cognitron_vl//third_party/ModelLink/
Looking in links: /data/software/
Obtaining file://local_disk/cognitron_vl/third_party/ModelLink
  Preparing metadata (setup.py): started
  Preparing metadata (setup.py): finished with status 'done'
Requirement already satisfied: numpy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.24.4)
Processing data/software/transformers-4.43.2-py3-none-any.whl (from modellink==0.0.1)
Processing data/software/transformers-stream-generator-0.0.5.tar.gz (from modellink==0.0.1)
  Preparing metadata (setup.py): started
  Preparing metadata (setup.py): finished with status 'done'
Requirement already satisfied: sympy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.4)
Requirement already satisfied: decorator in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (5.1.1)
Requirement already satisfied: scipy in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.10.1)
Requirement already satisfied: sentencepiece in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.2.0)
Requirement already satisfied: einops in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.7.0)
Requirement already satisfied: datasets in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (3.0.0)
Requirement already satisfied: pybind11 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (2.13.6)
Requirement already satisfied: accelerate in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.34.2)
Requirement already satisfied: six in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (1.16.0)
Requirement already satisfied: protobuf in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (4.25.3)
Processing data/software/peft-0.7.1-py3-none-any.whl (from modellink==0.0.1)
Requirement already satisfied: tiktoken in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from modellink==0.0.1) (0.7.0)
Requirement already satisfied: packaging>=20.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (23.2)
Requirement already satisfied: psutil in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (5.9.8)
Requirement already satisfied: pyyaml in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (5.4.1)
Requirement already satisfied: torch>=1.13.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (2.1.0+cpu)
Requirement already satisfied: tqdm in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (4.67.1)
Requirement already satisfied: safetensors in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (0.4.5)
Requirement already satisfied: huggingface-hub>=0.17.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from peft==0.7.1->modellink==0.0.1) (0.26.2)
Requirement already satisfied: filelock in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (3.13.1)
Requirement already satisfied: regex!=2019.12.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (2023.12.25)
Requirement already satisfied: requests in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from transformers==4.43.2->modellink==0.0.1) (2.32.3)
Processing data/software/tokenizers-0.19.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (from transformers==4.43.2->modellink==0.0.1)
Requirement already satisfied: pyarrow>=15.0.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (17.0.0)
Requirement already satisfied: dill<0.3.9,>=0.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (0.3.7)
Requirement already satisfied: pandas in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (2.0.3)
Requirement already satisfied: xxhash in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (3.4.1)
Requirement already satisfied: multiprocess in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (0.70.15)
Requirement already satisfied: fsspec<=2024.6.1,>=2023.1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from fsspec[http]<=2024.6.1,>=2023.1.0->datasets->modellink==0.0.1) (2023.10.0)
Requirement already satisfied: aiohttp in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from datasets->modellink==0.0.1) (3.9.3)
Requirement already satisfied: mpmath>=0.19 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from sympy->modellink==0.0.1) (1.3.0)
Requirement already satisfied: aiosignal>=1.1.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.3.1)
Requirement already satisfied: attrs>=17.3.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (23.2.0)
Requirement already satisfied: frozenlist>=1.1.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.4.1)
Requirement already satisfied: multidict<7.0,>=4.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (6.0.5)
Requirement already satisfied: yarl<2.0,>=1.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (1.9.4)
Requirement already satisfied: async-timeout<5.0,>=4.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from aiohttp->datasets->modellink==0.0.1) (4.0.3)
Requirement already satisfied: typing-extensions>=3.7.4.3 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from huggingface-hub>=0.17.0->peft==0.7.1->modellink==0.0.1) (4.10.0)
Requirement already satisfied: charset-normalizer<4,>=2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (3.3.2)
Requirement already satisfied: idna<4,>=2.5 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (3.6)
Requirement already satisfied: urllib3<3,>=1.21.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (1.26.18)
Requirement already satisfied: certifi>=2017.4.17 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from requests->transformers==4.43.2->modellink==0.0.1) (2024.2.2)
Requirement already satisfied: networkx in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (3.1)
Requirement already satisfied: jinja2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (3.1.3)
Requirement already satisfied: python-dateutil>=2.8.2 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2.8.2)
Requirement already satisfied: pytz>=2020.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2024.1)
Requirement already satisfied: tzdata>=2022.1 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from pandas->datasets->modellink==0.0.1) (2024.1)
Requirement already satisfied: MarkupSafe>=2.0 in /root/miniconda3/envs/py38/lib/python3.8/site-packages (from jinja2->torch>=1.13.0->peft==0.7.1->modellink==0.0.1) (2.1.5)
Building wheels for collected packages: transformers_stream_generator
  Building wheel for transformers_stream_generator (setup.py): started
  Building wheel for transformers_stream_generator (setup.py): finished with status 'done'
  Created wheel for transformers_stream_generator: filename=transformers_stream_generator-0.0.5-py3-none-any.whl size=12425 sha256=7b79cc70e3e25aee266f004baeb02f49a26365cbf13fc2e075375f124a308963
  Stored in directory: /root/.cache/pip/wheels/56/8c/42/5381d9c36bc85f28982f4cf8f98dc44d37a6d6c04897a5cb7c
Successfully built transformers_stream_generator
DEPRECATION: apex 0.1-ascend-20240523 has a non-standard version number. pip 24.1 will enforce this behaviour change. A possible replacement is to upgrade to a newer version of apex or contact the author to suggest that they release a version with a conforming version number. Discussion can be found at https://github.com/pypa/pip/issues/12063
Installing collected packages: tokenizers, transformers, transformers_stream_generator, peft, modellink
  Attempting uninstall: tokenizers
    Found existing installation: tokenizers 0.20.3
    Uninstalling tokenizers-0.20.3:
      Successfully uninstalled tokenizers-0.20.3
  Attempting uninstall: transformers
    Found existing installation: transformers 4.46.3
    Uninstalling transformers-4.46.3:
      Successfully uninstalled transformers-4.46.3
  Attempting uninstall: peft
    Found existing installation: peft 0.7.0
    Uninstalling peft-0.7.0:
      Successfully uninstalled peft-0.7.0
  Running setup.py develop for modellink
Successfully installed modellink-0.0.1 peft-0.7.1 tokenizers-0.19.1 transformers-4.43.2 transformers_stream_generator-0.0.5
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
+ export PYTHONPATH=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0//:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:
+ PYTHONPATH=/local_disk/cognitron_vl//third_party/Megatron-LM_core_r0.6.0//:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:/usr/local/Ascend/ascend-toolkit/latest/python/site-packages:/usr/local/Ascend/ascend-toolkit/latest/opp/built-in/op_impl/ai_core/tbe:
+ GPUS_PER_NODE=16
+ NNODES=32
+ NODE_RANK=7
+ MASTER_PORT=34567
+ export CUDA_DEVICE_MAX_CONNECTIONS=1
+ CUDA_DEVICE_MAX_CONNECTIONS=1
+ export PYTORCH_NPU_ALLOC_CONF=expandable_segments:True
+ PYTORCH_NPU_ALLOC_CONF=expandable_segments:True
+ VISION_SEQ_LENGTH=1025
+ IMAGE_TOKEN_LENGTH=256
+ IMAGE_SIZE=448
+ VISION_MODEL_TYPE=intern_300m
+ TP=8
+ PP=1
+ CP=8
+ CP_ALGO=megatron_cp_algo
+ CP_MASK=causal
+ DISTRIBUTED_ARGS='
    --nproc_per_node 16     --nnodes 32     --node_rank 7     --master_addr train-1198772881325351168-93vlj4s2getc-master-0.train-100034032793.svc.cluster.local     --master_port 34567
'
+ GPT_ARGS='
    --use-mcore-models     --tensor-model-parallel-size 8     --pipeline-model-parallel-size 1     --context-parallel-size 8     --context-parallel-algo megatron_cp_algo     --cp-attention-mask-type causal     --use-cp-send-recv-overlap     --no-create-attention-mask-in-dataloader     --sparse-mode 4     --sequence-parallel     --recompute-method block     --recompute-granularity full     --recompute-num-layers 48     --num-layers 48     --hidden-size 5120     --ffn-hidden-size 13824     --num-attention-heads 40     --group-query-attention     --num-query-groups 8     --tokenizer-type PretrainedFromHF     --tokenizer-name-or-path /data_4/models/Qwen/Qwen2.5-14B-Instruct/     --seq-length 1048576     --max-position-embeddings 1048576     --micro-batch-size 1     --global-batch-size 8     --make-vocab-size-divisible-by 1     --padded-vocab-size 152064     --rotary-base 1000000.0     --lr 5.00e-6     --train-iters 500     --lr-decay-style cosine     --untie-embeddings-and-output-weights     --disable-bias-linear     --attention-dropout 0.0     --init-method-std 0.01     --hidden-dropout 0.0     --position-embedding-type rope     --normalization RMSNorm     --use-fused-rmsnorm     --norm-epsilon 1e-6     --swiglu     --use-flash-attn     --use-fused-rotary-pos-emb     --use-rotary-position-embeddings     --use-fused-swiglu     --use-mc2     --no-masked-softmax-fusion     --attention-softmax-in-fp32     --min-lr 1.00e-7     --weight-decay 0.0     --lr-warmup-fraction 0.03     --clip-grad 1.0     --adam-beta1 0.9     --adam-beta2 0.999     --add-qkv-bias     --initial-loss-scale 4096     --no-gradient-accumulation-fusion     --use-distributed-optimizer     --bf16     --overlap-grad-reduce     --finetune     --vision-model-freeze     --vision-model-type intern_300m     --vision-downsample-ratio 0.5     --vision-projector-type mlp     --vision-projector-pre-norm     --vision-process-type dynamic     --vision-normalize-type imagenet     --vision-seq-length 1025     --image-token-length 256     --image-size 448     --prompt-format qwen2     --is-instruction-dataset     --max-num-image 4096     --max-fps 1     --add-class-token     --min-patch-grid 1     --max-patch-grid 12     --logit-mask     --cross-dataset-joint '
+ DATA_ARGS='
    --data-path /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml     --split 100,0,0     --data-seq-length 1048576     --num-workers 8 '
+ CKPT_ARGS='
    --load /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/     --vit-load /     --no-load-optim     --no-load-rng     --seed 42424242     --save /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743// '
+ OUTPUT_ARGS='
    --log-interval 1     --save-interval 20     --eval-interval 20     --eval-iters 0     --log-throughput     --distributed-timeout-minutes 120 '
+ torchrun --nproc_per_node 16 --nnodes 32 --node_rank 7 --master_addr train-1198772881325351168-93vlj4s2getc-master-0.train-100034032793.svc.cluster.local --master_port 34567 /local_disk/cognitron_vl//lcvlm_modellink/pretrain_lcvlm.py --use-mcore-models --tensor-model-parallel-size 8 --pipeline-model-parallel-size 1 --context-parallel-size 8 --context-parallel-algo megatron_cp_algo --cp-attention-mask-type causal --use-cp-send-recv-overlap --no-create-attention-mask-in-dataloader --sparse-mode 4 --sequence-parallel --recompute-method block --recompute-granularity full --recompute-num-layers 48 --num-layers 48 --hidden-size 5120 --ffn-hidden-size 13824 --num-attention-heads 40 --group-query-attention --num-query-groups 8 --tokenizer-type PretrainedFromHF --tokenizer-name-or-path /data_4/models/Qwen/Qwen2.5-14B-Instruct/ --seq-length 1048576 --max-position-embeddings 1048576 --micro-batch-size 1 --global-batch-size 8 --make-vocab-size-divisible-by 1 --padded-vocab-size 152064 --rotary-base 1000000.0 --lr 5.00e-6 --train-iters 500 --lr-decay-style cosine --untie-embeddings-and-output-weights --disable-bias-linear --attention-dropout 0.0 --init-method-std 0.01 --hidden-dropout 0.0 --position-embedding-type rope --normalization RMSNorm --use-fused-rmsnorm --norm-epsilon 1e-6 --swiglu --use-flash-attn --use-fused-rotary-pos-emb --use-rotary-position-embeddings --use-fused-swiglu --use-mc2 --no-masked-softmax-fusion --attention-softmax-in-fp32 --min-lr 1.00e-7 --weight-decay 0.0 --lr-warmup-fraction 0.03 --clip-grad 1.0 --adam-beta1 0.9 --adam-beta2 0.999 --add-qkv-bias --initial-loss-scale 4096 --no-gradient-accumulation-fusion --use-distributed-optimizer --bf16 --overlap-grad-reduce --finetune --vision-model-freeze --vision-model-type intern_300m --vision-downsample-ratio 0.5 --vision-projector-type mlp --vision-projector-pre-norm --vision-process-type dynamic --vision-normalize-type imagenet --vision-seq-length 1025 --image-token-length 256 --image-size 448 --prompt-format qwen2 --is-instruction-dataset --max-num-image 4096 --max-fps 1 --add-class-token --min-patch-grid 1 --max-patch-grid 12 --logit-mask --cross-dataset-joint --data-path /local_disk/cognitron_vl//configs/lcvlm_finetune_stage4.yaml --split 100,0,0 --data-seq-length 1048576 --num-workers 8 --load /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/ --vit-load / --no-load-optim --no-load-rng --seed 42424242 --save /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp8_stage4.sh/20241128_234743// --log-interval 1 --save-interval 20 --eval-interval 20 --eval-iters 0 --log-throughput --distributed-timeout-minutes 120 --distributed-backend nccl
[2024-11-28 15:50:27,310] torch.distributed.run: [WARNING] 
[2024-11-28 15:50:27,310] torch.distributed.run: [WARNING] *****************************************
[2024-11-28 15:50:27,310] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. 
[2024-11-28 15:50:27,310] torch.distributed.run: [WARNING] *****************************************
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Creating extension directory /root/.cache/torch_extensions/py38_cpu/adaptive_cp...
Emitting ninja build file /root/.cache/torch_extensions/py38_cpu/adaptive_cp/build.ninja...
Building extension module adaptive_cp...
Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N)
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
Using /root/.cache/torch_extensions/py38_cpu as PyTorch extensions root...
[1/2] c++ -MMD -MF adaptive_cp.o.d -DTORCH_EXTENSION_NAME=adaptive_cp -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE=\"_gcc\" -DPYBIND11_STDLIB=\"_libstdcpp\" -DPYBIND11_BUILD_ABI=\"_cxxabi1011\" -I/usr/local/Ascend/ascend-toolkit/latest/include -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/include -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/third_party -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/acl -I/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/inc -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/TH -isystem /root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/include/THC -isystem /root/miniconda3/envs/py38/include/python3.8 -D_GLIBCXX_USE_CXX11_ABI=0 -fPIC -std=c++17 -fstack-protector-all -Wl,-z,relro,-z,now,-z,noexecstack -fPIC -pie -Wl,--disable-new-dtags,--rpath -s -O2 -c local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/ops/csrc/algorithm/adaptive_cp/adaptive_cp.cpp -o adaptive_cp.o 
[2/2] c++ adaptive_cp.o -shared -L/usr/local/Ascend/ascend-toolkit/latest/lib64 -lascendcl -L/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch_npu/lib -ltorch_npu -L/root/miniconda3/envs/py38/lib/python3.8/site-packages/torch/lib -lc10 -ltorch_cpu -ltorch -ltorch_python -o adaptive_cp.so
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
Loading extension module adaptive_cp...
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
Loading extension module adaptive_cp...
local_disk/cognitron_vl/third_party/MindSpeed_core_r0.6.0/mindspeed/core/tensor_parallel/layers.py:30: UserWarning: failed to generate the npu_matmul_add_fp32
  warnings.warn("failed to generate the npu_matmul_add_fp32")
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
/root/miniconda3/envs/py38/lib/python3.8/site-packages/torchvision/io/image.py:13: UserWarning: Failed to load image Python extension: 'libc10_cuda.so: cannot open shared object file: No such file or directory'If you don't plan on using image functionality from `torchvision.io`, you can ignore this warning. Otherwise, there might be something wrong with your environment. Did you have `libjpeg` or `libpng` installed before building `torchvision` from source?
  warn(
> compiling dataset index builder ...
make: Entering directory 'local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0/megatron/core/datasets'
make: Nothing to be done for 'default'.
make: Leaving directory 'local_disk/cognitron_vl/third_party/Megatron-LM_core_r0.6.0/megatron/core/datasets'
>>> done with dataset index builder. Compilation time: 1.011 seconds
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute Falsevision_projector_recompute Falsevision_projector_recompute False


vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_projector_recompute False
vision_model_freeze
=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.vision_model_freeze

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.

=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

vision_model_freeze
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.vision_model_freeze

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

vision_model_freeze=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
vision_model_freeze=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.





=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
vision_model_freeze=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.




vision_model_freeze=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.





=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.




=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.






=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.






=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.





=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.






=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.






=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.





=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.






=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.





=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.vision_model_freeze=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.
vision_model_freeze
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

vision_model_freeze=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

vision_model_freeze=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.





=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.vision_model_freeze=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.vision_model_freeze
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.vision_model_freeze


=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.






=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.





=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.





=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.




=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.



=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.


=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.

=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
vision_model_freeze
=> set param external_feature_model.vit.class_token torch.Size([1, 1, 1024]) requires grad to False.
=> set param external_feature_model.vit.conv1.weight torch.Size([1024, 3, 14, 14]) requires grad to False.
=> set param external_feature_model.vit.conv1.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.position_embeddings.weight torch.Size([1025, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.0.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.1.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.2.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.3.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.4.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.5.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.6.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.7.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.8.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.9.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.10.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.11.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.12.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.13.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.14.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.15.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.16.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.17.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.18.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.19.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.20.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.21.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.22.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls1 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.ls2 torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.input_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.weight torch.Size([1024, 128]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_proj.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.weight torch.Size([384, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.self_attention.linear_qkv.bias torch.Size([384]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.weight torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.pre_mlp_layernorm.bias torch.Size([1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.weight torch.Size([512, 1024]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc1.bias torch.Size([512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.weight torch.Size([1024, 512]) requires grad to False.
=> set param external_feature_model.vit.decoder.layers.23.mlp.linear_fc2.bias torch.Size([1024]) requires grad to False.
model GPTVLModel(
  (external_feature_model): MegatronVisionModel(
    (vit): InternViTModel(
      (conv1): Conv2d(3, 1024, kernel_size=(14, 14), stride=(14, 14))
      (position_embeddings): Embedding(1025, 1024)
      (decoder): TransformerBlock(
        (layers): ModuleList(
          (0-23): 24 x InternViTTransformerLayer(
            (input_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (self_attention): SelfAttention(
              (core_attention): DotProductAttention(
                (scale_mask_softmax): FusedScaleMaskSoftmax()
                (attention_dropout): Dropout(p=0.0, inplace=False)
              )
              (linear_proj): RowParallelLinear()
              (linear_qkv): ColumnParallelLinear()
            )
            (self_attn_bda): IdentityFuncOp()
            (pre_cross_attn_layernorm): IdentityOp()
            (cross_attention): IdentityOp()
            (cross_attn_bda): IdentityFuncOp()
            (pre_mlp_layernorm): LayerNorm((1024,), eps=1e-06, elementwise_affine=True)
            (mlp): MLP(
              (linear_fc1): ColumnParallelLinear()
              (linear_fc2): RowParallelLinear()
            )
            (mlp_bda): IdentityFuncOp()
          )
        )
      )
    )
    (vision_projection): MultimodalProjector(
      (encoder): MLP(
        (linear_fc1): ColumnParallelLinear()
        (linear_fc2): RowParallelLinear()
      )
    )
    (pre_proj_layernorm): LayerNorm((4096,), eps=1e-05, elementwise_affine=True)
  )
  (embedding): LanguageModelEmbedding(
    (word_embeddings): VocabParallelEmbedding()
    (embedding_dropout): Dropout(p=0.0, inplace=False)
  )
  (rotary_pos_emb): RotaryEmbedding()
  (decoder): TransformerBlock(
    (layers): ModuleList(
      (0-47): 48 x TransformerLayer(
        (input_layernorm): RMSNorm()
        (self_attention): SelfAttention(
          (core_attention): DotProductAttention(
            (scale_mask_softmax): FusedScaleMaskSoftmax()
            (attention_dropout): Dropout(p=0.0, inplace=False)
          )
          (linear_proj): RowParallelLinear()
          (linear_qkv): ColumnParallelLinear()
          (q_layernorm): IdentityOp()
          (k_layernorm): IdentityOp()
        )
        (pre_cross_attn_layernorm): IdentityOp()
        (cross_attention): IdentityOp()
        (cross_attn_bda): IdentityFuncOp()
        (pre_mlp_layernorm): RMSNorm()
        (mlp): MLP(
          (linear_fc1): ColumnParallelLinear()
          (linear_fc2): RowParallelLinear()
        )
      )
    )
    (final_layernorm): RMSNorm()
  )
  (output_layer): ColumnParallelLinear()
)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)





_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)






_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)




_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)





_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)





_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)





_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)






_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)





_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)






_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.vision_projection.encoder.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.external_feature_model.pre_proj_layernorm.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.embedding.word_embeddings.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.0.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.0.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.1.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.1.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.2.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.2.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.2.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.3.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)







_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.3.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.3.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.4.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.5.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.5.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.6.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.6.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.7.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.7.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.7.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.8.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.8.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.9.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.9.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.9.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.10.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.10.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.10.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.11.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.self_attention.linear_proj.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.12.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.13.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.13.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.14.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.15.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.15.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.16.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.16.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.16.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)




_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.17.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.17.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.18.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.18.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.19.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)



_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.19.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.20.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.20.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.21.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.21.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.22.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.22.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.22.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc1.weight key (1.0, 1.0, False, False)


_get_param_groups name module.module.decoder.layers.22.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.23.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.23.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.24.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.24.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.25.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.25.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.26.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.26.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.27.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.27.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.28.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.29.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.29.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.30.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.31.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.31.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.32.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.32.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.33.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.34.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.35.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.36.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.36.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.37.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.37.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.38.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.39.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.39.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.40.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.40.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.input_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.41.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.41.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.42.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.42.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.43.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc1.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.43.mlp.linear_fc2.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)_get_param_groups name module.module.decoder.layers.44.self_attention.linear_proj.weight key (1.0, 1.0, False, False)

_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.44.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.45.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.46.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.input_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_proj.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.self_attention.linear_qkv.bias key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.pre_mlp_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc1.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.layers.47.mlp.linear_fc2.weight key (1.0, 1.0, False, False)
_get_param_groups name module.module.decoder.final_layernorm.weight key (0.0, 1.0, False, False)
_get_param_groups name module.module.output_layer.weight key (1.0, 1.0, False, False)
_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration _load_base_checkpoint iteration_load_base_checkpoint iteration_load_base_checkpoint iteration _load_base_checkpoint iteration _load_base_checkpoint iteration  1000 1000  
1000_load_base_checkpoint iteration _load_base_checkpoint release  100010001000

1000 1000_load_base_checkpoint iteration1000
 1000 1000 10001000

_load_base_checkpoint release_load_base_checkpoint release_load_base_checkpoint release   

_load_base_checkpoint release

1000
False_load_base_checkpoint release

  _load_base_checkpoint release_load_base_checkpoint release FalseFalse

_load_base_checkpoint release_load_base_checkpoint release 1000False _load_base_checkpoint release_load_base_checkpoint release 
 False
_load_base_checkpoint releaseFalse 
1000  False
 
False
_load_base_checkpoint releaseFalse
FalseFalse


FalseFalse
 

False_load_base_checkpoint release_load_base_checkpoint release
  
FalseFalse

_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_07/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_05/model_optim_rng.pt
_load_base_checkpoint_load_base_checkpoint  /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_04/model_optim_rng.pt/data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_04/model_optim_rng.pt

_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_06/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_02/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_07/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_00/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_05/model_optim_rng.pt_load_base_checkpoint
 /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_02/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_01/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_03/model_optim_rng.pt_load_base_checkpoint
 /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_03/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_01/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_06/model_optim_rng.pt
_load_base_checkpoint /data_2/output/LM/scripts/modellink/qwen25/finetune_qwen25_14b_intern_300m_ptd_tp8pp1cp2_stage3.sh/20241127_204213/iter_0001000/mp_rank_00/model_optim_rng.pt
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
load_checkpoint iteration 0
load_checkpoint release False
strict True
> rank 120 is creating GPT datasets ...> rank 127 does not create GPT datasets ...

> rank 122 does not create GPT datasets ...
> rank 119 does not create GPT datasets ...
> rank 121 does not create GPT datasets ...
> rank 117 does not create GPT datasets ...
> rank 123 does not create GPT datasets ...
> rank 112 is creating GPT datasets ...
> rank 125 does not create GPT datasets ...
> rank 115 does not create GPT datasets ...
> rank 114 does not create GPT datasets ...
> rank 126 does not create GPT datasets ...
> rank 116 does not create GPT datasets ...
> rank 124 does not create GPT datasets ...
> rank 118 does not create GPT datasets ...
> rank 113 does not create GPT datasets ...
target_ratios [(1, 1), (1, 2), (2, 1), (3, 1), (1, 3), (2, 2), (4, 1), (1, 4), (5, 1), (1, 5), (1, 6), (6, 1), (3, 2), (2, 3), (7, 1), (1, 7), (4, 2), (2, 4), (1, 8), (8, 1), (1, 9), (3, 3), (9, 1), (2, 5), (5, 2), (10, 1), (1, 10), (11, 1), (1, 11), (12, 1), (3, 4), (4, 3), (1, 12), (6, 2), (2, 6)]
possible_resolutions [[448, 448], [448, 896], [896, 448], [1344, 448], [448, 1344], [896, 896], [1792, 448], [448, 1792], [2240, 448], [448, 2240], [448, 2688], [2688, 448], [1344, 896], [896, 1344], [3136, 448], [448, 3136], [1792, 896], [896, 1792], [448, 3584], [3584, 448], [448, 4032], [1344, 1344], [4032, 448], [896, 2240], [2240, 896], [4480, 448], [448, 4480], [4928, 448], [448, 4928], [5376, 448], [1344, 1792], [1792, 1344], [448, 5376], [2688, 896], [896, 2688]]
target_ratios [(1, 1), (1, 2), (2, 1), (3, 1), (1, 3), (2, 2), (4, 1), (1, 4), (5, 1), (1, 5), (1, 6), (6, 1), (3, 2), (2, 3), (7, 1), (1, 7), (4, 2), (2, 4), (1, 8), (8, 1), (1, 9), (3, 3), (9, 1), (2, 5), (5, 2), (10, 1), (1, 10), (11, 1), (1, 11), (12, 1), (3, 4), (4, 3), (1, 12), (6, 2), (2, 6)]
possible_resolutions [[448, 448], [448, 896], [896, 448], [1344, 448], [448, 1344], [896, 896], [1792, 448], [448, 1792], [2240, 448], [448, 2240], [448, 2688], [2688, 448], [1344, 896], [896, 1344], [3136, 448], [448, 3136], [1792, 896], [896, 1792], [448, 3584], [3584, 448], [448, 4032], [1344, 1344], [4032, 448], [896, 2240], [2240, 896], [4480, 448], [448, 4480], [4928, 448], [448, 4928], [5376, 448], [1344, 1792], [1792, 1344], [448, 5376], [2688, 896], [896, 2688]]
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a4b72f80] mmco: unref short failure
[h264 @ 0x5592a4b72f80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a237c940] mmco: unref short failure
[h264 @ 0x5592a237c940] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a237c940] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
processed_samples 100 unjoint_samples 100 joint_samples 0 [148295, 146271]
processed_samples 100 unjoint_samples 100 joint_samples 0 [148295, 146271]
processed_samples 100 unjoint_samples 100 joint_samples 0 [226774, 195274]
processed_samples 100 unjoint_samples 100 joint_samples 0 [226774, 195274]
processed_samples 100 unjoint_samples 100 joint_samples 0 [151573, 140769]
processed_samples 100 unjoint_samples 100 joint_samples 0 [151573, 140769]
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
processed_samples 100 unjoint_samples 100 joint_samples 0 [159460, 158169]
processed_samples 100 unjoint_samples 100 joint_samples 0 [159460, 158169]
processed_samples 100 unjoint_samples 100 joint_samples 0 [114507, 113006]
processed_samples 100 unjoint_samples 100 joint_samples 0 [114507, 113006]
processed_samples 100 unjoint_samples 100 joint_samples 0 [206494, 206498]
processed_samples 100 unjoint_samples 100 joint_samples 0 [206494, 206498]
processed_samples 100 unjoint_samples 100 joint_samples 0 [127988, 140214]
processed_samples 100 unjoint_samples 100 joint_samples 0 [127988, 140214]
processed_samples 100 unjoint_samples 100 joint_samples 0 [214923, 223827]
processed_samples 100 unjoint_samples 100 joint_samples 0 [214923, 223827]
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97aa0b980] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a1bfcec0] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
processed_samples 200 unjoint_samples 200 joint_samples 0 [310024, 316961]
processed_samples 200 unjoint_samples 200 joint_samples 0 [338165, 327792]
processed_samples 200 unjoint_samples 200 joint_samples 0 [338165, 327792]
processed_samples 200 unjoint_samples 200 joint_samples 0 [310024, 316961]
processed_samples 200 unjoint_samples 200 joint_samples 0 [306060, 306437]
processed_samples 200 unjoint_samples 200 joint_samples 0 [306060, 306437]
processed_samples 200 unjoint_samples 200 joint_samples 0 [362739, 364102]
processed_samples 200 unjoint_samples 200 joint_samples 0 [362739, 364102]
processed_samples 200 unjoint_samples 200 joint_samples 0 [246837, 246769]
processed_samples 200 unjoint_samples 200 joint_samples 0 [246837, 246769]
processed_samples 200 unjoint_samples 200 joint_samples 0 [372269, 396147]
processed_samples 200 unjoint_samples 200 joint_samples 0 [372269, 396147]
processed_samples 200 unjoint_samples 200 joint_samples 0 [383369, 384977]
processed_samples 200 unjoint_samples 200 joint_samples 0 [383369, 384977]
processed_samples 200 unjoint_samples 200 joint_samples 0 [260780, 258622]
processed_samples 200 unjoint_samples 200 joint_samples 0 [260780, 258622]
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] [h264 @ 0x55d97dd14f40] mmco: unref short failure
mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
processed_samples 300 unjoint_samples 300 joint_samples 0 [534967, 536276]
processed_samples 300 unjoint_samples 300 joint_samples 0 [534967, 536276]
processed_samples 300 unjoint_samples 300 joint_samples 0 [632262, 542552]
processed_samples 300 unjoint_samples 300 joint_samples 0 [539739, 538279]
processed_samples 300 unjoint_samples 300 joint_samples 0 [632262, 542552]
processed_samples 300 unjoint_samples 300 joint_samples 0 [539739, 538279]
processed_samples 300 unjoint_samples 300 joint_samples 0 [481971, 480342]
processed_samples 300 unjoint_samples 300 joint_samples 0 [481971, 480342]
processed_samples 300 unjoint_samples 300 joint_samples 0 [472163, 470544]
processed_samples 300 unjoint_samples 300 joint_samples 0 [472163, 470544]
processed_samples 300 unjoint_samples 300 joint_samples 0 [383926, 380134]
processed_samples 300 unjoint_samples 300 joint_samples 0 [383926, 380134]
processed_samples 300 unjoint_samples 300 joint_samples 0 [427140, 461956]
processed_samples 300 unjoint_samples 300 joint_samples 0 [427140, 461956]
processed_samples 300 unjoint_samples 300 joint_samples 0 [434357, 433640]
processed_samples 300 unjoint_samples 300 joint_samples 0 [434357, 433640]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97facaf00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 400 unjoint_samples 400 joint_samples 0 [632939, 632939]
processed_samples 400 unjoint_samples 400 joint_samples 0 [584307, 589060]
processed_samples 400 unjoint_samples 400 joint_samples 0 [584307, 589060]
[h264 @ 0x55d97ac67100] mmco: unref short failure
processed_samples 400 unjoint_samples 400 joint_samples 0 [632939, 632939]
[h264 @ 0x5592a0bf6640] mmco: unref short failure
processed_samples 400 unjoint_samples 400 joint_samples 0 [630240, 577556]
processed_samples 400 unjoint_samples 400 joint_samples 0 [656659, 657895]
processed_samples 400 unjoint_samples 400 joint_samples 0 [803209, 826234]
processed_samples 400 unjoint_samples 400 joint_samples 0 [803209, 826234]
processed_samples 400 unjoint_samples 400 joint_samples 0 [656659, 657895]
processed_samples 400 unjoint_samples 400 joint_samples 0 [520109, 498185]
processed_samples 400 unjoint_samples 400 joint_samples 0 [520109, 498185]
processed_samples 400 unjoint_samples 400 joint_samples 0 [630240, 577556]
processed_samples 400 unjoint_samples 400 joint_samples 0 [621127, 622592]
processed_samples 400 unjoint_samples 400 joint_samples 0 [621127, 622592]
processed_samples 400 unjoint_samples 400 joint_samples 0 [698946, 696652]
processed_samples 400 unjoint_samples 400 joint_samples 0 [698946, 696652]
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
processed_samples 500 unjoint_samples 500 joint_samples 0 [712926, 714246]
processed_samples 500 unjoint_samples 500 joint_samples 0 [712926, 714246]
processed_samples 500 unjoint_samples 500 joint_samples 0 [839495, 863168]
processed_samples 500 unjoint_samples 500 joint_samples 0 [839495, 863168]
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
processed_samples 500 unjoint_samples 500 joint_samples 0 [831603, 834681]
processed_samples 500 unjoint_samples 500 joint_samples 0 [831603, 834681]
processed_samples 500 unjoint_samples 500 joint_samples 0 [777295, 774851]
processed_samples 500 unjoint_samples 500 joint_samples 0 [777295, 774851]
processed_samples 500 unjoint_samples 500 joint_samples 0 [942194, 940015]
processed_samples 500 unjoint_samples 500 joint_samples 0 [942194, 940015]
processed_samples 500 unjoint_samples 500 joint_samples 0 [734875, 818819]
processed_samples 500 unjoint_samples 500 joint_samples 0 [734875, 818819]
processed_samples 500 unjoint_samples 500 joint_samples 0 [786212, 782871]
processed_samples 500 unjoint_samples 500 joint_samples 0 [786212, 782871]
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
processed_samples 500 unjoint_samples 500 joint_samples 0 [830756, 830609]
processed_samples 500 unjoint_samples 500 joint_samples 0 [830756, 830609]
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x5592a1a7eb00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
................................................................................................[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 600 unjoint_samples 600 joint_samples 0 [983365, 978422]
processed_samples 600 unjoint_samples 600 joint_samples 0 [983365, 978422]
processed_samples 600 unjoint_samples 600 joint_samples 1 [61011, 1046651]
processed_samples 600 unjoint_samples 600 joint_samples 1 [61011, 1046651]
processed_samples 600 unjoint_samples 600 joint_samples 0 [972070, 968484]
processed_samples 600 unjoint_samples 600 joint_samples 0 [981506, 979579]
processed_samples 600 unjoint_samples 600 joint_samples 0 [972070, 968484]
processed_samples 600 unjoint_samples 600 joint_samples 0 [981506, 979579]
processed_samples 600 unjoint_samples 600 joint_samples 0 [906832, 928256]
processed_samples 600 unjoint_samples 600 joint_samples 0 [906832, 928256]
processed_samples 600 unjoint_samples 600 joint_samples 0 [941812, 955487]
processed_samples 600 unjoint_samples 600 joint_samples 0 [941812, 955487]
processed_samples 600 unjoint_samples 600 joint_samples 0 [899766, 898832]
processed_samples 600 unjoint_samples 600 joint_samples 0 [899766, 898832]
processed_samples 600 unjoint_samples 600 joint_samples 0 [924524, 923930]
processed_samples 600 unjoint_samples 600 joint_samples 0 [924524, 923930]
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
processed_samples 700 unjoint_samples 700 joint_samples 1 [1047584, 131968]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1047584, 131968]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1034552, 154838]
processed_samples 700 unjoint_samples 700 joint_samples 1 [997187, 171575]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1034552, 154838]
processed_samples 700 unjoint_samples 700 joint_samples 1 [997187, 171575]
processed_samples 700 unjoint_samples 700 joint_samples 1 [411057, 1046651]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1046489, 116149]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1046489, 116149]
processed_samples 700 unjoint_samples 700 joint_samples 1 [411057, 1046651]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1048173, 162694]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1048173, 162694]
processed_samples 700 unjoint_samples 700 joint_samples 1 [181107, 1041274]
processed_samples 700 unjoint_samples 700 joint_samples 1 [181107, 1041274]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1040753, 168643]
processed_samples 700 unjoint_samples 700 joint_samples 1 [1040753, 168643]
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97facaf00] mmco: unref short failure
[h264 @ 0x55d97facaf00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a1bfcec0] mmco: unref short failure
[h264 @ 0x5592a1bfcec0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
processed_samples 800 unjoint_samples 800 joint_samples 1 [1034552, 474672]
processed_samples 800 unjoint_samples 800 joint_samples 1 [1034552, 474672]
processed_samples 800 unjoint_samples 800 joint_samples 1 [502540, 1041274]
processed_samples 800 unjoint_samples 800 joint_samples 1 [502540, 1041274]
processed_samples 800 unjoint_samples 800 joint_samples 1 [1046489, 462747]
processed_samples 800 unjoint_samples 800 joint_samples 1 [1046489, 462747]
processed_samples 800 unjoint_samples 800 joint_samples 1 [1047584, 513407]
processed_samples 800 unjoint_samples 800 joint_samples 1 [1047584, 513407]
processed_samples 800 unjoint_samples 800 joint_samples 1 [1048173, 469703]
processed_samples 800 unjoint_samples 800 joint_samples 1 [1048173, 469703]
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 800 unjoint_samples 800 joint_samples 1 [997187, 415269]
processed_samples 800 unjoint_samples 800 joint_samples 1 [997187, 415269]
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
processed_samples 800 unjoint_samples 800 joint_samples 1 [1040753, 447491]
processed_samples 800 unjoint_samples 800 joint_samples 1 [1040753, 447491]
processed_samples 800 unjoint_samples 800 joint_samples 1 [698219, 1046651]
processed_samples 800 unjoint_samples 800 joint_samples 1 [698219, 1046651]
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a446ee80] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
processed_samples 900 unjoint_samples 900 joint_samples 1 [1046489, 753769]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1046489, 753769]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1047584, 769497]
processed_samples 900 unjoint_samples 900 joint_samples 1 [997187, 868616]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1034552, 776272]
processed_samples 900 unjoint_samples 900 joint_samples 1 [862675, 1041274]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1047584, 769497]
processed_samples 900 unjoint_samples 900 joint_samples 1 [997187, 868616]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1034552, 776272]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1048173, 849280]
processed_samples 900 unjoint_samples 900 joint_samples 1 [862675, 1041274]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1048173, 849280]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1040753, 823881]
processed_samples 900 unjoint_samples 900 joint_samples 1 [1040753, 823881]
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
processed_samples 900 unjoint_samples 900 joint_samples 1 [995252, 1046651]
processed_samples 900 unjoint_samples 900 joint_samples 1 [995252, 1046651]
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a466b500] mmco: unref short failure
[h264 @ 0x5592a466b500] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1046489, 25577]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1046489, 25577]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1047584, 9023]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1047584, 9023]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [245420, 1046651]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [245420, 1046651]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1047023, 129024]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1047023, 129024]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [152537, 1024639]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1041938, 170346]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [152537, 1024639]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [1041938, 170346]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [59426, 1027395]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [59426, 1027395]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [160010, 1047501]
processed_samples 1000 unjoint_samples 1000 joint_samples 2 [160010, 1047501]
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a466b500] mmco: unref short failure
[h264 @ 0x5592a466b500] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1046489, 456997]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1046489, 456997]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [431031, 1024639]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [431031, 1024639]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1047584, 260249]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1047584, 260249]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [452232, 1047501]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [452232, 1047501]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1047023, 470365]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1047023, 470365]
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [281062, 1027395]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [281062, 1027395]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [523478, 1046651]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [523478, 1046651]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1041938, 477841]
processed_samples 1100 unjoint_samples 1100 joint_samples 2 [1041938, 477841]
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1047584, 486021]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1047584, 486021]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [701937, 1047501]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [701937, 1047501]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1041938, 780591]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1041938, 780591]
[h264 @ 0x5592a1bfcec0] mmco: unref short failure
[h264 @ 0x5592a1bfcec0] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [715717, 1024639]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [715717, 1024639]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1046489, 820933]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1046489, 820933]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [547344, 1027395]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [547344, 1027395]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [826811, 1046651]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [826811, 1046651]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1047023, 802644]
processed_samples 1200 unjoint_samples 1200 joint_samples 2 [1047023, 802644]
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97eb65780] mmco: unref short failure
[h264 @ 0x55d97eb65780] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97eb65780] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x5592a22642c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [11959, 1046561]
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [11959, 1046561]
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [1046489, 153637]
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [1046489, 153637]
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [26500, 1046481]
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [26500, 1046481]
processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1047584, 710856]
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [109243, 1046190]
processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1047584, 710856]
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [68126, 1046651]
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [109243, 1046190]
processed_samples 1300 unjoint_samples 1300 joint_samples 3 [68126, 1046651]
processed_samples 1300 unjoint_samples 1300 joint_samples 2 [926060, 1027395]
processed_samples 1300 unjoint_samples 1300 joint_samples 2 [926060, 1027395]
processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1017715, 1047501]
processed_samples 1300 unjoint_samples 1300 joint_samples 2 [1017715, 1047501]
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [234427, 1040688]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [234427, 1040688]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [316423, 1046481]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1047027, 126319]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [316423, 1046481]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1047027, 126319]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [308971, 1046651]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [308555, 1046561]
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [308971, 1046651]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [409459, 1046190]
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [308555, 1046561]
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [409459, 1046190]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1047306, 218873]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1047306, 218873]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1046489, 392507]
processed_samples 1400 unjoint_samples 1400 joint_samples 3 [1046489, 392507]
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x5592a0a3c780] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x55d97ad8ebc0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [538476, 1040688]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [538476, 1040688]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [680511, 1046481]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [680511, 1046481]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [618831, 1046561]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [618831, 1046561]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1047306, 615897]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1047306, 615897]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1047027, 557381]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1047027, 557381]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [662622, 1046651]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [662622, 1046651]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1046489, 710921]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [1046489, 710921]
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [778983, 1046190]
processed_samples 1500 unjoint_samples 1500 joint_samples 3 [778983, 1046190]
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a1c80f80] co located POCs unavailable
[h264 @ 0x55d97b490480] co located POCs unavailable
[h264 @ 0x55d97b490480] co located POCs unavailable
[h264 @ 0x5592a1c80f80] co located POCs unavailable
[h264 @ 0x55d97b490480] co located POCs unavailable
[h264 @ 0x5592a1c80f80] co located POCs unavailable
processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1036930, 22963]
processed_samples 1600 unjoint_samples 1600 joint_samples 4 [1036930, 22963]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [859914, 1040688]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [859914, 1040688]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1028306, 1046481]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1028306, 1046481]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1047306, 970175]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1047306, 970175]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1047027, 912560]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [919811, 1046651]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [960149, 1046561]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [960149, 1046561]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1047027, 912560]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [919811, 1046651]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1046489, 979581]
processed_samples 1600 unjoint_samples 1600 joint_samples 3 [1046489, 979581]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x55d97e5c4c40] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a13cdd80] mmco: unref short failure
[h264 @ 0x5592a13cdd80] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x5592a1559240] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1036836, 157070]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1036836, 157070]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1007009, 429433]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1007009, 429433]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [210643, 1046675]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [210643, 1046675]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1036930, 327451]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1036930, 327451]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1047027, 102894]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1033979, 310342]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1047027, 102894]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [1033979, 310342]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [141671, 1046651]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [141671, 1046651]
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [222980, 1046634]
processed_samples 1700 unjoint_samples 1700 joint_samples 4 [222980, 1046634]
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [419794, 1046675]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [419794, 1046675]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1036836, 550904]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1036836, 550904]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1033979, 698982]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [523557, 1046651]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [523557, 1046651]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1033979, 698982]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1007009, 738011]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1007009, 738011]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1036930, 598447]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1036930, 598447]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [599687, 1046634]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [599687, 1046634]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1047027, 348612]
processed_samples 1800 unjoint_samples 1800 joint_samples 4 [1047027, 348612]
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x55d97eb71e80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a08edc00] [h264 @ 0x55d97a987100] mmco: unref short failure
mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97e018040] mmco: unref short failure
[h264 @ 0x55d97e018040] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
processed_samples 1900 unjoint_samples 1900 joint_samples 5 [61822, 1016920]
processed_samples 1900 unjoint_samples 1900 joint_samples 5 [61822, 1016920]
processed_samples 1900 unjoint_samples 1900 joint_samples 5 [73251, 1022454]
processed_samples 1900 unjoint_samples 1900 joint_samples 5 [73251, 1022454]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [946435, 1046634]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1047100, 1047274]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [946435, 1046634]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1047100, 1047274]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1047027, 585234]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1047027, 585234]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1036836, 758733]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [1036836, 758733]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [772695, 1046651]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [772695, 1046651]
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [702819, 1046675]
[h264 @ 0x5592a4a4c140] mmco: unref short failure
processed_samples 1900 unjoint_samples 1900 joint_samples 4 [702819, 1046675]
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 2000 unjoint_samples 2000 joint_samples 5 [315775, 1016920]
processed_samples 2000 unjoint_samples 2000 joint_samples 5 [315775, 1016920]
processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1047100, 282520]
processed_samples 2000 unjoint_samples 2000 joint_samples 5 [1047100, 282520]
processed_samples 2000 unjoint_samples 2000 joint_samples 5 [339034, 1022454]
processed_samples 2000 unjoint_samples 2000 joint_samples 4 [993636, 1046675]
processed_samples 2000 unjoint_samples 2000 joint_samples 4 [993636, 1046675]
processed_samples 2000 unjoint_samples 2000 joint_samples 5 [339034, 1022454]
processed_samples 2000 unjoint_samples 2000 joint_samples 5 [203783, 1046634]
processed_samples 2000 unjoint_samples 2000 joint_samples 5 [203783, 1046634]
processed_samples 2000 unjoint_samples 2000 joint_samples 4 [1036836, 1019277]
processed_samples 2000 unjoint_samples 2000 joint_samples 4 [1040258, 1046651]
processed_samples 2000 unjoint_samples 2000 joint_samples 4 [1036836, 1019277]
processed_samples 2000 unjoint_samples 2000 joint_samples 4 [1040258, 1046651]
processed_samples 2000 unjoint_samples 2000 joint_samples 4 [1047027, 1006664]
processed_samples 2000 unjoint_samples 2000 joint_samples 4 [1047027, 1006664]
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a0c9d140] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a23746c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1047100, 597957]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1047100, 597957]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [443976, 1046634]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1019502, 266617]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [575333, 1016920]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [443976, 1046634]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1019502, 266617]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [575333, 1016920]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1044266, 385402]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [1044266, 385402]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [336254, 1046651]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [336254, 1046651]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [391213, 1032016]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [391213, 1032016]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [696188, 1022454]
processed_samples 2100 unjoint_samples 2100 joint_samples 5 [696188, 1022454]
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
processed_samples 2200 unjoint_samples 2200 joint_samples 6 [1046183, 25196]
processed_samples 2200 unjoint_samples 2200 joint_samples 6 [1046183, 25196]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1044266, 721448]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1044266, 721448]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [670397, 1032016]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [670397, 1032016]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [705241, 1046634]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [705241, 1046634]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1047100, 937359]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1047100, 937359]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1019502, 557629]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [1019502, 557629]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [908055, 1016920]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [908055, 1016920]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [627312, 1046651]
processed_samples 2200 unjoint_samples 2200 joint_samples 5 [627312, 1046651]
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
processed_samples 2300 unjoint_samples 2300 joint_samples 6 [20509, 1048364]
processed_samples 2300 unjoint_samples 2300 joint_samples 6 [20509, 1048364]
processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1047100, 198025]
processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1047100, 198025]
processed_samples 2300 unjoint_samples 2300 joint_samples 6 [157155, 1037330]
processed_samples 2300 unjoint_samples 2300 joint_samples 6 [157155, 1037330]
processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1046183, 302065]
processed_samples 2300 unjoint_samples 2300 joint_samples 6 [1046183, 302065]
processed_samples 2300 unjoint_samples 2300 joint_samples 5 [1019502, 837851]
processed_samples 2300 unjoint_samples 2300 joint_samples 5 [1019502, 837851]
processed_samples 2300 unjoint_samples 2300 joint_samples 5 [937169, 1046651]
processed_samples 2300 unjoint_samples 2300 joint_samples 5 [937169, 1046651]
processed_samples 2300 unjoint_samples 2300 joint_samples 5 [948451, 1032016]
processed_samples 2300 unjoint_samples 2300 joint_samples 5 [948451, 1032016]
processed_samples 2300 unjoint_samples 2300 joint_samples 5 [984890, 1046634]
processed_samples 2300 unjoint_samples 2300 joint_samples 5 [984890, 1046634]
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [371682, 1048364]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [371682, 1048364]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1023424, 253397]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1023424, 253397]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [365371, 1037330]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [365371, 1037330]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1047100, 640133]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1047100, 640133]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1046391, 141736]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1046391, 141736]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [203653, 1038002]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [203653, 1038002]
processed_samples 2400 unjoint_samples 2400 joint_samples 5 [1043835, 1046064]
processed_samples 2400 unjoint_samples 2400 joint_samples 5 [1043835, 1046064]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1046183, 615609]
processed_samples 2400 unjoint_samples 2400 joint_samples 6 [1046183, 615609]
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1023424, 561190]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1023424, 561190]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1045931, 279838]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1045931, 279838]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [631582, 1037330]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [637317, 1048364]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [631582, 1037330]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [637317, 1048364]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1047100, 991182]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1047100, 991182]
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1046391, 407351]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1046391, 407351]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1046183, 1016017]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [1046183, 1016017]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [536355, 1038002]
processed_samples 2500 unjoint_samples 2500 joint_samples 6 [536355, 1038002]
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
processed_samples 2600 unjoint_samples 2600 joint_samples 7 [1047100, 320970]
processed_samples 2600 unjoint_samples 2600 joint_samples 7 [1047100, 320970]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1045931, 686535]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1045931, 686535]
processed_samples 2600 unjoint_samples 2600 joint_samples 7 [378250, 1045550]
processed_samples 2600 unjoint_samples 2600 joint_samples 7 [378250, 1045550]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1046391, 714649]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [953699, 1037330]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1046391, 714649]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [953699, 1037330]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1007627, 1048364]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1007627, 1048364]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1023424, 824807]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [805253, 1038002]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [1023424, 824807]
processed_samples 2600 unjoint_samples 2600 joint_samples 6 [805253, 1038002]
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a3659c80] mmco: unref short failure
[h264 @ 0x5592a3659c80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a3659c80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x55d97e40bf80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [43842, 1043492]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [43842, 1043492]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1039431, 260760]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [78573, 1042991]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1039431, 260760]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [78573, 1042991]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1039278, 199808]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1039278, 199808]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1047100, 670322]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [1047100, 670322]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [724212, 1045550]
processed_samples 2700 unjoint_samples 2700 joint_samples 7 [724212, 1045550]
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
processed_samples 2700 unjoint_samples 2700 joint_samples 6 [1046391, 985647]
processed_samples 2700 unjoint_samples 2700 joint_samples 6 [1046391, 985647]
processed_samples 2700 unjoint_samples 2700 joint_samples 6 [1045931, 984540]
processed_samples 2700 unjoint_samples 2700 joint_samples 6 [1045931, 984540]
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a6114740] mmco: unref short failure
[h264 @ 0x5592a6114740] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 2800 unjoint_samples 2800 joint_samples 8 [1047100, 29082]
processed_samples 2800 unjoint_samples 2800 joint_samples 8 [1047100, 29082]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [501351, 1042991]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [501351, 1042991]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1039431, 538364]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1039431, 538364]
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1046439, 189839]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1046439, 189839]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1047934, 312538]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1047934, 312538]
processed_samples 2800 unjoint_samples 2800 joint_samples 8 [1038359, 46686]
processed_samples 2800 unjoint_samples 2800 joint_samples 8 [1038359, 46686]
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1039278, 530984]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [1039278, 530984]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [321453, 1043492]
processed_samples 2800 unjoint_samples 2800 joint_samples 7 [321453, 1043492]
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a6114740] mmco: unref short failure
[h264 @ 0x5592a6114740] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97facaf00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1047100, 409875]
processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1047100, 409875]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1047934, 623274]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [776160, 1042991]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [776160, 1042991]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1047934, 623274]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1039431, 769562]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1039431, 769562]
processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1038359, 332162]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [585878, 1043492]
processed_samples 2900 unjoint_samples 2900 joint_samples 8 [1038359, 332162]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [585878, 1043492]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1046439, 492469]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1046439, 492469]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1039278, 902668]
processed_samples 2900 unjoint_samples 2900 joint_samples 7 [1039278, 902668]
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a6114740] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1005120, 68573]
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1005120, 68573]
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1043347, 153392]
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [34667, 1046472]
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1043347, 153392]
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [34667, 1046472]
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1047100, 685955]
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1047100, 685955]
processed_samples 3000 unjoint_samples 3000 joint_samples 7 [1046439, 811628]
processed_samples 3000 unjoint_samples 3000 joint_samples 7 [1046439, 811628]
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1038359, 665245]
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
processed_samples 3000 unjoint_samples 3000 joint_samples 8 [1038359, 665245]
processed_samples 3000 unjoint_samples 3000 joint_samples 7 [1047934, 969223]
processed_samples 3000 unjoint_samples 3000 joint_samples 7 [1047934, 969223]
processed_samples 3000 unjoint_samples 3000 joint_samples 7 [854909, 1043492]
processed_samples 3000 unjoint_samples 3000 joint_samples 7 [854909, 1043492]
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a1adfa80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [296119, 1046472]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [296119, 1046472]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [69211, 1047145]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1045647, 182033]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [69211, 1047145]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1045647, 182033]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1005120, 362586]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1005120, 362586]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1043347, 457602]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1043347, 457602]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1047100, 1037484]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1047100, 1037484]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [287290, 1036458]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [287290, 1036458]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1038359, 922120]
processed_samples 3100 unjoint_samples 3100 joint_samples 8 [1038359, 922120]
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e7305c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
processed_samples 3200 unjoint_samples 3200 joint_samples 9 [290338, 1047050]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [580382, 1036458]
processed_samples 3200 unjoint_samples 3200 joint_samples 9 [290338, 1047050]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [553609, 1046472]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1043347, 803472]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [553609, 1046472]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [580382, 1036458]
processed_samples 3200 unjoint_samples 3200 joint_samples 9 [1047501, 201421]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1045647, 417007]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1043347, 803472]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [340904, 1047145]
processed_samples 3200 unjoint_samples 3200 joint_samples 9 [1047501, 201421]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1045647, 417007]
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [340904, 1047145]
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1005120, 764934]
processed_samples 3200 unjoint_samples 3200 joint_samples 8 [1005120, 764934]
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e7305c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
processed_samples 3300 unjoint_samples 3300 joint_samples 9 [566079, 1047050]
processed_samples 3300 unjoint_samples 3300 joint_samples 9 [566079, 1047050]
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [808367, 1036458]
processed_samples 3300 unjoint_samples 3300 joint_samples 9 [136201, 1047071]
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [808367, 1036458]
processed_samples 3300 unjoint_samples 3300 joint_samples 9 [136201, 1047071]
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [566161, 1047145]
processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1047501, 503577]
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [566161, 1047145]
processed_samples 3300 unjoint_samples 3300 joint_samples 9 [1047501, 503577]
[h264 @ 0x5592a0445b00] mmco: unref short failure
[h264 @ 0x5592a0445b00] mmco: unref short failure
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [898251, 1046472]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [898251, 1046472]
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [1045647, 733558]
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [1045647, 733558]
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [1007417, 1006476]
processed_samples 3300 unjoint_samples 3300 joint_samples 8 [1007417, 1006476]
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a50afd00] mmco: unref short failure
[h264 @ 0x5592a50afd00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [437028, 1047071]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [437028, 1047071]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1047255, 272827]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1047255, 272827]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [166682, 1046472]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [166682, 1046472]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [78530, 1044867]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [78530, 1044867]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1003158, 1047050]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1003158, 1047050]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1047501, 896356]
processed_samples 3400 unjoint_samples 3400 joint_samples 9 [1047501, 896356]
processed_samples 3400 unjoint_samples 3400 joint_samples 8 [942106, 1047145]
processed_samples 3400 unjoint_samples 3400 joint_samples 8 [942106, 1047145]
processed_samples 3400 unjoint_samples 3400 joint_samples 8 [1045647, 1035817]
processed_samples 3400 unjoint_samples 3400 joint_samples 8 [1045647, 1035817]
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0f8c100] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [494911, 1044867]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [494911, 1044867]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1047255, 568897]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1047255, 568897]
processed_samples 3500 unjoint_samples 3500 joint_samples 10 [1036571, 219146]
processed_samples 3500 unjoint_samples 3500 joint_samples 10 [1036571, 219146]
processed_samples 3500 unjoint_samples 3500 joint_samples 10 [1047501, 163063]
processed_samples 3500 unjoint_samples 3500 joint_samples 10 [1047501, 163063]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1045647, 308665]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [1045647, 308665]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [221854, 1047145]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [221854, 1047145]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [772238, 1047071]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [772238, 1047071]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [577120, 1046472]
processed_samples 3500 unjoint_samples 3500 joint_samples 9 [577120, 1046472]
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a5093a80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
processed_samples 3600 unjoint_samples 3600 joint_samples 10 [9584, 1047334]
processed_samples 3600 unjoint_samples 3600 joint_samples 10 [9584, 1047334]
processed_samples 3600 unjoint_samples 3600 joint_samples 10 [1036571, 512453]
processed_samples 3600 unjoint_samples 3600 joint_samples 10 [1036571, 512453]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [817355, 1044867]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [817355, 1044867]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [1045647, 634702]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [1045647, 634702]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [504836, 1047145]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [504836, 1047145]
processed_samples 3600 unjoint_samples 3600 joint_samples 10 [1047501, 540815]
processed_samples 3600 unjoint_samples 3600 joint_samples 10 [1047501, 540815]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [904064, 1046472]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [904064, 1046472]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [1047255, 901706]
processed_samples 3600 unjoint_samples 3600 joint_samples 9 [1047255, 901706]
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a47279c0] mmco: unref short failure
[h264 @ 0x5592a47279c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1046555, 175842]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1047255, 299584]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1046555, 175842]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [305737, 1047334]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1047255, 299584]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [305737, 1047334]
processed_samples 3700 unjoint_samples 3700 joint_samples 9 [828606, 1047145]
processed_samples 3700 unjoint_samples 3700 joint_samples 9 [828606, 1047145]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1047501, 844890]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1026451, 102170]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1026451, 102170]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1047501, 844890]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1036571, 846159]
processed_samples 3700 unjoint_samples 3700 joint_samples 10 [1036571, 846159]
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
processed_samples 3700 unjoint_samples 3700 joint_samples 9 [1045647, 878876]
processed_samples 3700 unjoint_samples 3700 joint_samples 9 [1045647, 878876]
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d9815ff9c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
processed_samples 3800 unjoint_samples 3800 joint_samples 11 [1046508, 54842]
processed_samples 3800 unjoint_samples 3800 joint_samples 11 [1046508, 54842]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1046553, 72421]
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1046553, 72421]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1047871, 135768]
processed_samples 3800 unjoint_samples 3800 joint_samples 11 [1047501, 113162]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1047871, 135768]
processed_samples 3800 unjoint_samples 3800 joint_samples 11 [1047501, 113162]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1046555, 485136]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1046555, 485136]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1026451, 343429]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1026451, 343429]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1047255, 529113]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [1047255, 529113]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [615270, 1047334]
processed_samples 3800 unjoint_samples 3800 joint_samples 10 [615270, 1047334]
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1708a80] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1046508, 449792]
processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1046508, 449792]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1046553, 336929]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1046553, 336929]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1046555, 758127]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1046555, 758127]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1047255, 848201]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1047255, 848201]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1047871, 425619]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1047871, 425619]
processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1025990, 23489]
processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1025990, 23489]
processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1047501, 396463]
processed_samples 3900 unjoint_samples 3900 joint_samples 11 [1047501, 396463]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1026451, 712066]
processed_samples 3900 unjoint_samples 3900 joint_samples 10 [1026451, 712066]
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [92348, 1044954]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [92348, 1044954]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1025990, 243533]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1025990, 243533]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1046508, 788163]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1046508, 788163]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [232686, 1014030]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [232686, 1014030]
processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1047871, 712115]
processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1047871, 712115]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1047501, 676551]
processed_samples 4000 unjoint_samples 4000 joint_samples 11 [1047501, 676551]
processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1046553, 787511]
processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1046553, 787511]
processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1026451, 977922]
processed_samples 4000 unjoint_samples 4000 joint_samples 10 [1026451, 977922]
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1708a80] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [394583, 1044954]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [394583, 1044954]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [271526, 1038351]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [271526, 1038351]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [546361, 1014030]
processed_samples 4100 unjoint_samples 4100 joint_samples 10 [1046553, 1023351]
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [546361, 1014030]
processed_samples 4100 unjoint_samples 4100 joint_samples 10 [1046553, 1023351]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1025990, 544622]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1025990, 544622]
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
processed_samples 4100 unjoint_samples 4100 joint_samples 10 [1047871, 1020212]
processed_samples 4100 unjoint_samples 4100 joint_samples 10 [1047871, 1020212]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1046508, 1039703]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1046508, 1039703]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1047501, 1038758]
processed_samples 4100 unjoint_samples 4100 joint_samples 11 [1047501, 1038758]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1708a80] mmco: unref short failure
[h264 @ 0x5592a1708a80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97acc1a00] mmco: unref short failure
[h264 @ 0x55d97acc1a00] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [523617, 1038351]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [523617, 1038351]
processed_samples 4200 unjoint_samples 4200 joint_samples 12 [240241, 1047824]
processed_samples 4200 unjoint_samples 4200 joint_samples 12 [240241, 1047824]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [661469, 1044954]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [661469, 1044954]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [301802, 1037517]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [348231, 1040028]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [348231, 1040028]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [301802, 1037517]
processed_samples 4200 unjoint_samples 4200 joint_samples 12 [273013, 1041046]
processed_samples 4200 unjoint_samples 4200 joint_samples 12 [273013, 1041046]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1025990, 841836]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [1025990, 841836]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [897590, 1014030]
processed_samples 4200 unjoint_samples 4200 joint_samples 11 [897590, 1014030]
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1708a80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
processed_samples 4300 unjoint_samples 4300 joint_samples 12 [681205, 1047824]
processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1046624, 132602]
processed_samples 4300 unjoint_samples 4300 joint_samples 12 [255192, 1040664]
processed_samples 4300 unjoint_samples 4300 joint_samples 12 [513614, 1041046]
processed_samples 4300 unjoint_samples 4300 joint_samples 11 [678979, 1040028]
processed_samples 4300 unjoint_samples 4300 joint_samples 11 [978860, 1044954]
processed_samples 4300 unjoint_samples 4300 joint_samples 11 [583238, 1037517]
processed_samples 4300 unjoint_samples 4300 joint_samples 12 [681205, 1047824]
processed_samples 4300 unjoint_samples 4300 joint_samples 12 [1046624, 132602]
processed_samples 4300 unjoint_samples 4300 joint_samples 12 [255192, 1040664]
processed_samples 4300 unjoint_samples 4300 joint_samples 12 [513614, 1041046]
processed_samples 4300 unjoint_samples 4300 joint_samples 11 [678979, 1040028]
processed_samples 4300 unjoint_samples 4300 joint_samples 11 [583238, 1037517]
processed_samples 4300 unjoint_samples 4300 joint_samples 11 [978860, 1044954]
processed_samples 4300 unjoint_samples 4300 joint_samples 11 [834006, 1038351]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 4300 unjoint_samples 4300 joint_samples 11 [834006, 1038351]
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a13abe80] mmco: unref short failure
[h264 @ 0x55d983320600] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
processed_samples 4400 unjoint_samples 4400 joint_samples 13 [1047355, 38192]
processed_samples 4400 unjoint_samples 4400 joint_samples 13 [1047355, 38192]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [180399, 1046066]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [180399, 1046066]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1006622, 1041046]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1006622, 1041046]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1046624, 626305]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1046624, 626305]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [559301, 1040664]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [559301, 1040664]
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [51898, 1047650]
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [51898, 1047650]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1045907, 95801]
processed_samples 4400 unjoint_samples 4400 joint_samples 12 [1045907, 95801]
processed_samples 4400 unjoint_samples 4400 joint_samples 11 [903789, 1037517]
processed_samples 4400 unjoint_samples 4400 joint_samples 11 [903789, 1037517]
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] Missing reference picture, default is 65524
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] Missing reference picture, default is 65524
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a5905c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a5905c0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b282700] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b282700] mmco: unref short failure
[h264 @ 0x55d97b282700] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97b282700] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b282700] mmco: unref short failure
[h264 @ 0x55d97b282700] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1045907, 470791]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1045907, 470791]
processed_samples 4500 unjoint_samples 4500 joint_samples 13 [1047355, 271072]
processed_samples 4500 unjoint_samples 4500 joint_samples 13 [1047355, 271072]
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [917173, 1040664]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [917173, 1040664]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [70001, 1043461]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [70001, 1043461]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [579883, 1046066]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [579883, 1046066]
processed_samples 4500 unjoint_samples 4500 joint_samples 13 [1047047, 340321]
processed_samples 4500 unjoint_samples 4500 joint_samples 13 [1047047, 340321]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [454808, 1047650]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [454808, 1047650]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1046624, 894842]
processed_samples 4500 unjoint_samples 4500 joint_samples 12 [1046624, 894842]
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b0c5880] mmco: unref short failure
[h264 @ 0x55d97b0c5880] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b0c5880] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1047355, 582140]
processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1047355, 582140]
processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1042180, 192992]
processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1042180, 192992]
processed_samples 4600 unjoint_samples 4600 joint_samples 13 [187540, 1002183]
processed_samples 4600 unjoint_samples 4600 joint_samples 13 [187540, 1002183]
processed_samples 4600 unjoint_samples 4600 joint_samples 12 [950411, 1046066]
processed_samples 4600 unjoint_samples 4600 joint_samples 12 [950411, 1046066]
processed_samples 4600 unjoint_samples 4600 joint_samples 12 [306905, 1043461]
processed_samples 4600 unjoint_samples 4600 joint_samples 12 [306905, 1043461]
processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1047047, 656409]
processed_samples 4600 unjoint_samples 4600 joint_samples 13 [1047047, 656409]
processed_samples 4600 unjoint_samples 4600 joint_samples 12 [1045907, 782146]
processed_samples 4600 unjoint_samples 4600 joint_samples 12 [1045907, 782146]
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
processed_samples 4600 unjoint_samples 4600 joint_samples 12 [998724, 1047650]
processed_samples 4600 unjoint_samples 4600 joint_samples 12 [998724, 1047650]
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a4a64840] mmco: unref short failure
[h264 @ 0x5592a4a64840] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1046178, 188393]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1046178, 188393]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [432666, 1002183]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [432666, 1002183]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1042180, 559665]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1042180, 559665]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047357, 293198]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047357, 293198]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047355, 904527]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047355, 904527]
processed_samples 4701 unjoint_samples 4700 joint_samples 12 [731412, 1043461]
processed_samples 4701 unjoint_samples 4700 joint_samples 12 [731412, 1043461]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047047, 935782]
processed_samples 4700 unjoint_samples 4700 joint_samples 13 [1047047, 935782]
processed_samples 4700 unjoint_samples 4700 joint_samples 12 [1045907, 1026923]
processed_samples 4700 unjoint_samples 4700 joint_samples 12 [1045907, 1026923]
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e6c8e00] mmco: unref short failure
[h264 @ 0x55d97e6c8e00] mmco: unref short failure
processed_samples 4800 unjoint_samples 4800 joint_samples 14 [226329, 1036602]
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1046178, 413892]
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [319936, 1026923]
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [758853, 1002183]
processed_samples 4800 unjoint_samples 4800 joint_samples 14 [1047047, 179412]
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1047357, 578385]
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [319936, 1026923]
processed_samples 4800 unjoint_samples 4800 joint_samples 14 [226329, 1036602]
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1046178, 413892]
processed_samples 4801 unjoint_samples 4800 joint_samples 12 [974709, 1043461]
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [758853, 1002183]
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1042180, 973520]
processed_samples 4800 unjoint_samples 4800 joint_samples 14 [1047047, 179412]
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1047357, 578385]
processed_samples 4801 unjoint_samples 4800 joint_samples 12 [974709, 1043461]
[h264 @ 0x55d97be2e400] mmco: unref short failure
processed_samples 4800 unjoint_samples 4800 joint_samples 13 [1042180, 973520]
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a15cab80] mmco: unref short failure
[h264 @ 0x5592a15cab80] mmco: unref short failure
[h264 @ 0x5592a15cab80] mmco: unref short failure
[h264 @ 0x5592a15cab80] mmco: unref short failure
[h264 @ 0x5592a15cab80] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a5109340] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1046178, 716739]
processed_samples 4900 unjoint_samples 4900 joint_samples 14 [501029, 1036602]
processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1046178, 716739]
processed_samples 4900 unjoint_samples 4900 joint_samples 14 [501029, 1036602]
processed_samples 4900 unjoint_samples 4900 joint_samples 13 [957258, 1002183]
processed_samples 4900 unjoint_samples 4900 joint_samples 14 [269131, 1033002]
processed_samples 4900 unjoint_samples 4900 joint_samples 13 [957258, 1002183]
processed_samples 4900 unjoint_samples 4900 joint_samples 14 [269131, 1033002]
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
processed_samples 4901 unjoint_samples 4900 joint_samples 13 [236026, 1048193]
processed_samples 4900 unjoint_samples 4900 joint_samples 14 [1047047, 438839]
processed_samples 4900 unjoint_samples 4900 joint_samples 13 [654542, 1026923]
processed_samples 4901 unjoint_samples 4900 joint_samples 13 [236026, 1048193]
processed_samples 4900 unjoint_samples 4900 joint_samples 14 [1047047, 438839]
[h264 @ 0x55d979dbe500] mmco: unref short failure
processed_samples 4900 unjoint_samples 4900 joint_samples 13 [654542, 1026923]
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1047357, 875888]
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
processed_samples 4900 unjoint_samples 4900 joint_samples 13 [1047357, 875888]
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a5109340] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
processed_samples 5000 unjoint_samples 5000 joint_samples 13 [1046178, 989753]
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [792609, 1036602]
[h264 @ 0x5592a4e100c0] mmco: unref short failure
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [335031, 1021688]
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [208035, 1046542]
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
processed_samples 5000 unjoint_samples 5000 joint_samples 13 [1046178, 989753]
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [208035, 1046542]
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [792609, 1036602]
[h264 @ 0x55d97e64a980] mmco: unref short failure
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [335031, 1021688]
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [696529, 1033002]
processed_samples 5000 unjoint_samples 5000 joint_samples 13 [940205, 1026923]
processed_samples 5000 unjoint_samples 5000 joint_samples 13 [940205, 1026923]
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [696529, 1033002]
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1047047, 815354]
processed_samples 5001 unjoint_samples 5000 joint_samples 13 [736520, 1048193]
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
processed_samples 5000 unjoint_samples 5000 joint_samples 14 [1047047, 815354]
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 5001 unjoint_samples 5000 joint_samples 13 [736520, 1048193]
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a4e6e000] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a5a07f80] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
processed_samples 5100 unjoint_samples 5100 joint_samples 15 [106938, 1038720]
processed_samples 5100 unjoint_samples 5100 joint_samples 15 [106938, 1038720]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [141786, 1046985]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [141786, 1046985]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [249455, 1046407]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [249455, 1046407]
processed_samples 5100 unjoint_samples 5100 joint_samples 15 [1048251, 84586]
processed_samples 5100 unjoint_samples 5100 joint_samples 15 [1048251, 84586]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [607726, 1021688]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [607726, 1021688]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [577724, 1046542]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [577724, 1046542]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [997919, 1033002]
processed_samples 5100 unjoint_samples 5100 joint_samples 14 [997919, 1033002]
processed_samples 5101 unjoint_samples 5100 joint_samples 13 [1010794, 1048193]
processed_samples 5101 unjoint_samples 5100 joint_samples 13 [1010794, 1048193]
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97eb1c340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a4e6e000] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97c72f5c0] mmco: unref short failure
[h264 @ 0x55d97c72f5c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97c72f5c0] mmco: unref short failure
[h264 @ 0x55d97c72f5c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1045951, 306352]
processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1045951, 306352]
processed_samples 5200 unjoint_samples 5200 joint_samples 15 [503810, 1038720]
processed_samples 5200 unjoint_samples 5200 joint_samples 15 [503810, 1038720]
processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1048251, 401520]
processed_samples 5200 unjoint_samples 5200 joint_samples 14 [593063, 1046985]
processed_samples 5200 unjoint_samples 5200 joint_samples 15 [1048251, 401520]
processed_samples 5200 unjoint_samples 5200 joint_samples 14 [593063, 1046985]
processed_samples 5200 unjoint_samples 5200 joint_samples 14 [609928, 1046407]
processed_samples 5200 unjoint_samples 5200 joint_samples 14 [609928, 1046407]
processed_samples 5201 unjoint_samples 5200 joint_samples 14 [1047562, 329810]
processed_samples 5201 unjoint_samples 5200 joint_samples 14 [1047562, 329810]
processed_samples 5200 unjoint_samples 5200 joint_samples 14 [953484, 1021688]
processed_samples 5200 unjoint_samples 5200 joint_samples 14 [953484, 1021688]
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
processed_samples 5200 unjoint_samples 5200 joint_samples 14 [836247, 1046542]
processed_samples 5200 unjoint_samples 5200 joint_samples 14 [836247, 1046542]
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a13cdd80] mmco: unref short failure
[h264 @ 0x5592a13cdd80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [97562, 1046542]
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1045951, 618201]
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [208570, 1036519]
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [837095, 1038720]
processed_samples 5300 unjoint_samples 5300 joint_samples 14 [886019, 1046985]
processed_samples 5301 unjoint_samples 5300 joint_samples 14 [1047562, 549029]
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [97562, 1046542]
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1048251, 731803]
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1045951, 618201]
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [208570, 1036519]
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [837095, 1038720]
processed_samples 5300 unjoint_samples 5300 joint_samples 14 [932120, 1046407]
processed_samples 5300 unjoint_samples 5300 joint_samples 14 [886019, 1046985]
processed_samples 5300 unjoint_samples 5300 joint_samples 15 [1048251, 731803]
processed_samples 5301 unjoint_samples 5300 joint_samples 14 [1047562, 549029]
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
processed_samples 5300 unjoint_samples 5300 joint_samples 14 [932120, 1046407]
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a1cd2400] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 5400 unjoint_samples 5400 joint_samples 16 [82474, 1042200]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [92234, 1046985]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [92234, 1046985]
processed_samples 5400 unjoint_samples 5400 joint_samples 16 [82474, 1042200]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [194807, 1046407]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [194807, 1046407]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1045951, 1030307]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1045951, 1030307]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [470763, 1036519]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [470763, 1036519]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1048251, 1031835]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [1048251, 1031835]
processed_samples 5401 unjoint_samples 5400 joint_samples 14 [1047562, 802293]
processed_samples 5401 unjoint_samples 5400 joint_samples 14 [1047562, 802293]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [433256, 1046542]
processed_samples 5400 unjoint_samples 5400 joint_samples 15 [433256, 1046542]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
processed_samples 5500 unjoint_samples 5500 joint_samples 16 [329756, 1042200]
processed_samples 5500 unjoint_samples 5500 joint_samples 15 [466376, 1046407]
processed_samples 5500 unjoint_samples 5500 joint_samples 16 [329756, 1042200]
processed_samples 5500 unjoint_samples 5500 joint_samples 15 [466376, 1046407]
processed_samples 5500 unjoint_samples 5500 joint_samples 15 [574705, 1046985]
processed_samples 5500 unjoint_samples 5500 joint_samples 15 [574705, 1046985]
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
processed_samples 5500 unjoint_samples 5500 joint_samples 16 [260476, 1046186]
processed_samples 5500 unjoint_samples 5500 joint_samples 16 [449452, 1046900]
processed_samples 5500 unjoint_samples 5500 joint_samples 16 [260476, 1046186]
processed_samples 5500 unjoint_samples 5500 joint_samples 16 [449452, 1046900]
processed_samples 5500 unjoint_samples 5500 joint_samples 15 [718680, 1036519]
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 5501 unjoint_samples 5500 joint_samples 15 [139142, 1028105]
processed_samples 5500 unjoint_samples 5500 joint_samples 15 [718680, 1036519]
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 5501 unjoint_samples 5500 joint_samples 15 [139142, 1028105]
processed_samples 5500 unjoint_samples 5500 joint_samples 15 [686332, 1046542]
processed_samples 5500 unjoint_samples 5500 joint_samples 15 [686332, 1046542]
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
processed_samples 5600 unjoint_samples 5600 joint_samples 16 [703087, 1042200]
processed_samples 5600 unjoint_samples 5600 joint_samples 16 [630759, 1046186]
processed_samples 5600 unjoint_samples 5600 joint_samples 15 [887128, 1046985]
processed_samples 5600 unjoint_samples 5600 joint_samples 16 [846124, 1046900]
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
processed_samples 5600 unjoint_samples 5600 joint_samples 15 [1046110, 1046542]
processed_samples 5600 unjoint_samples 5600 joint_samples 15 [811749, 1046407]
processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1031656, 44409]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 5600 unjoint_samples 5600 joint_samples 15 [887128, 1046985]
processed_samples 5600 unjoint_samples 5600 joint_samples 16 [630759, 1046186]
processed_samples 5600 unjoint_samples 5600 joint_samples 16 [846124, 1046900]
processed_samples 5600 unjoint_samples 5600 joint_samples 15 [811749, 1046407]
processed_samples 5600 unjoint_samples 5600 joint_samples 16 [703087, 1042200]
processed_samples 5600 unjoint_samples 5600 joint_samples 15 [1046110, 1046542]
processed_samples 5600 unjoint_samples 5600 joint_samples 16 [1031656, 44409]
processed_samples 5601 unjoint_samples 5600 joint_samples 15 [386501, 1028105]
processed_samples 5601 unjoint_samples 5600 joint_samples 15 [386501, 1028105]
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
processed_samples 5700 unjoint_samples 5700 joint_samples 17 [48371, 1047789]
processed_samples 5700 unjoint_samples 5700 joint_samples 17 [48371, 1047789]
processed_samples 5700 unjoint_samples 5700 joint_samples 17 [1040027, 29556]
processed_samples 5700 unjoint_samples 5700 joint_samples 17 [1040027, 29556]
processed_samples 5700 unjoint_samples 5700 joint_samples 17 [8472, 1047426]
processed_samples 5700 unjoint_samples 5700 joint_samples 17 [8472, 1047426]
processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1045231, 167207]
processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1045231, 167207]
processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1048434, 299548]
processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1032549, 124187]
processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1048434, 299548]
processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1032549, 124187]
processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1031656, 286474]
processed_samples 5700 unjoint_samples 5700 joint_samples 16 [1031656, 286474]
processed_samples 5701 unjoint_samples 5700 joint_samples 15 [682013, 1028105]
processed_samples 5701 unjoint_samples 5700 joint_samples 15 [682013, 1028105]
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a226e1c0] mmco: unref short failure
[h264 @ 0x5592a226e1c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1045231, 491063]
processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1032549, 440181]
processed_samples 5800 unjoint_samples 5800 joint_samples 17 [342192, 1047789]
processed_samples 5800 unjoint_samples 5800 joint_samples 17 [1040027, 288428]
processed_samples 5800 unjoint_samples 5800 joint_samples 17 [490296, 1047426]
processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1031656, 626860]
processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1048434, 553264]
processed_samples 5800 unjoint_samples 5800 joint_samples 17 [490296, 1047426]
[h264 @ 0x55d97a51b240] mmco: unref short failure
processed_samples 5800 unjoint_samples 5800 joint_samples 17 [1040027, 288428]
processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1045231, 491063]
processed_samples 5800 unjoint_samples 5800 joint_samples 17 [342192, 1047789]
processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1032549, 440181]
processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1031656, 626860]
processed_samples 5800 unjoint_samples 5800 joint_samples 16 [1048434, 553264]
processed_samples 5801 unjoint_samples 5800 joint_samples 15 [922687, 1028105]
processed_samples 5801 unjoint_samples 5800 joint_samples 15 [922687, 1028105]
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 5900 unjoint_samples 5900 joint_samples 17 [75328, 989920]
processed_samples 5900 unjoint_samples 5900 joint_samples 17 [75328, 989920]
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1032549, 770768]
processed_samples 5900 unjoint_samples 5900 joint_samples 17 [731916, 1047426]
processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1045231, 792920]
processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1045231, 792920]
processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1032549, 770768]
processed_samples 5900 unjoint_samples 5900 joint_samples 17 [731916, 1047426]
processed_samples 5901 unjoint_samples 5900 joint_samples 16 [147350, 1040054]
processed_samples 5901 unjoint_samples 5900 joint_samples 16 [147350, 1040054]
processed_samples 5900 unjoint_samples 5900 joint_samples 17 [677378, 1047789]
processed_samples 5900 unjoint_samples 5900 joint_samples 17 [677378, 1047789]
processed_samples 5900 unjoint_samples 5900 joint_samples 17 [1040027, 735195]
processed_samples 5900 unjoint_samples 5900 joint_samples 17 [1040027, 735195]
processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1048434, 806080]
processed_samples 5900 unjoint_samples 5900 joint_samples 16 [1048434, 806080]
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a2b5cac0] mmco: unref short failure
[h264 @ 0x5592a2b5cac0] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1045231, 47743]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1045231, 47743]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [101138, 1034327]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [101138, 1034327]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [66293, 1043610]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [66293, 1043610]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1040027, 1025456]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [359309, 989920]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [359309, 989920]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [998525, 1047426]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1040027, 1025456]
processed_samples 6001 unjoint_samples 6000 joint_samples 16 [456563, 1040054]
processed_samples 6001 unjoint_samples 6000 joint_samples 16 [456563, 1040054]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [998525, 1047426]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1010012, 1047789]
processed_samples 6000 unjoint_samples 6000 joint_samples 17 [1010012, 1047789]
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a0d867c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e6b5a80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
processed_samples 6100 unjoint_samples 6100 joint_samples 18 [1028759, 282189]
processed_samples 6100 unjoint_samples 6100 joint_samples 18 [1046447, 244648]
processed_samples 6100 unjoint_samples 6100 joint_samples 18 [1028759, 282189]
processed_samples 6100 unjoint_samples 6100 joint_samples 18 [239376, 1041483]
processed_samples 6100 unjoint_samples 6100 joint_samples 18 [1046447, 244648]
processed_samples 6100 unjoint_samples 6100 joint_samples 18 [239376, 1041483]
processed_samples 6100 unjoint_samples 6100 joint_samples 17 [363580, 1043610]
[h264 @ 0x55d97cd63980] mmco: unref short failure
processed_samples 6100 unjoint_samples 6100 joint_samples 17 [363580, 1043610]
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1045231, 364534]
processed_samples 6100 unjoint_samples 6100 joint_samples 17 [370227, 1034327]
processed_samples 6100 unjoint_samples 6100 joint_samples 17 [1045231, 364534]
processed_samples 6100 unjoint_samples 6100 joint_samples 17 [598403, 989920]
processed_samples 6100 unjoint_samples 6100 joint_samples 17 [370227, 1034327]
processed_samples 6100 unjoint_samples 6100 joint_samples 17 [598403, 989920]
processed_samples 6101 unjoint_samples 6100 joint_samples 16 [769512, 1040054]
processed_samples 6101 unjoint_samples 6100 joint_samples 16 [769512, 1040054]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97b40cc00] mmco: unref short failure
[h264 @ 0x55d97b40cc00] mmco: unref short failure
[h264 @ 0x55d97b40cc00] mmco: unref short failure
[h264 @ 0x55d97b40cc00] mmco: unref short failure
[h264 @ 0x55d97b40cc00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 6200 unjoint_samples 6200 joint_samples 18 [1046447, 611977]
processed_samples 6200 unjoint_samples 6200 joint_samples 17 [1045231, 626798]
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 6200 unjoint_samples 6200 joint_samples 18 [1028759, 601364]
processed_samples 6200 unjoint_samples 6200 joint_samples 18 [538611, 1041483]
processed_samples 6200 unjoint_samples 6200 joint_samples 17 [689154, 1043610]
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
processed_samples 6200 unjoint_samples 6200 joint_samples 17 [653786, 1034327]
processed_samples 6200 unjoint_samples 6200 joint_samples 17 [942382, 989920]
processed_samples 6201 unjoint_samples 6200 joint_samples 17 [1046849, 4507]
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
processed_samples 6200 unjoint_samples 6200 joint_samples 18 [1046447, 611977]
[h264 @ 0x5592a4e9a600] mmco: unref short failure
processed_samples 6200 unjoint_samples 6200 joint_samples 17 [689154, 1043610]
processed_samples 6200 unjoint_samples 6200 joint_samples 18 [1028759, 601364]
processed_samples 6200 unjoint_samples 6200 joint_samples 18 [538611, 1041483]
processed_samples 6200 unjoint_samples 6200 joint_samples 17 [1045231, 626798]
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
processed_samples 6200 unjoint_samples 6200 joint_samples 17 [942382, 989920]
processed_samples 6200 unjoint_samples 6200 joint_samples 17 [653786, 1034327]
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 6201 unjoint_samples 6200 joint_samples 17 [1046849, 4507]
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97b40cc00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97b40cc00] mmco: unref short failure
[h264 @ 0x55d97b40cc00] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2a46600] [h264 @ 0x55d97e5be6c0] mmco: unref short failure
mmco: unref short failure
[h264 @ 0x5592a2a46600] [h264 @ 0x55d97e5be6c0] mmco: unref short failure
mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
processed_samples 6300 unjoint_samples 6300 joint_samples 17 [1045231, 982086]
processed_samples 6300 unjoint_samples 6300 joint_samples 19 [38019, 1037049]
processed_samples 6300 unjoint_samples 6300 joint_samples 17 [1045231, 982086]
processed_samples 6300 unjoint_samples 6300 joint_samples 19 [38019, 1037049]
processed_samples 6300 unjoint_samples 6300 joint_samples 18 [978942, 148577]
processed_samples 6300 unjoint_samples 6300 joint_samples 18 [978942, 148577]
processed_samples 6300 unjoint_samples 6300 joint_samples 18 [1036519, 135445]
processed_samples 6300 unjoint_samples 6300 joint_samples 19 [22355, 1028150]
processed_samples 6300 unjoint_samples 6300 joint_samples 18 [1036519, 135445]
processed_samples 6300 unjoint_samples 6300 joint_samples 18 [805340, 1041483]
processed_samples 6300 unjoint_samples 6300 joint_samples 18 [805340, 1041483]
processed_samples 6301 unjoint_samples 6300 joint_samples 17 [1046849, 290663]
processed_samples 6300 unjoint_samples 6300 joint_samples 19 [22355, 1028150]
processed_samples 6301 unjoint_samples 6300 joint_samples 17 [1046849, 290663]
processed_samples 6300 unjoint_samples 6300 joint_samples 17 [1032801, 1043610]
processed_samples 6300 unjoint_samples 6300 joint_samples 17 [1032801, 1043610]
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97aea3b00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97aea3b00] mmco: unref short failure
[h264 @ 0x55d97aea3b00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97aea3b00] mmco: unref short failure
[h264 @ 0x55d97aea3b00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97aea3b00] mmco: unref short failure
[h264 @ 0x55d97aea3b00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
processed_samples 6400 unjoint_samples 6400 joint_samples 19 [1029273, 250240]
processed_samples 6400 unjoint_samples 6400 joint_samples 19 [1029273, 250240]
processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1046833, 284399]
processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1046833, 284399]
processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1036519, 381177]
processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1036519, 381177]
processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1043298, 407187]
processed_samples 6400 unjoint_samples 6400 joint_samples 18 [1043298, 407187]
processed_samples 6400 unjoint_samples 6400 joint_samples 19 [381713, 1037049]
processed_samples 6400 unjoint_samples 6400 joint_samples 19 [381713, 1037049]
processed_samples 6400 unjoint_samples 6400 joint_samples 18 [978942, 484096]
processed_samples 6400 unjoint_samples 6400 joint_samples 19 [291830, 1028150]
processed_samples 6400 unjoint_samples 6400 joint_samples 19 [291830, 1028150]
processed_samples 6400 unjoint_samples 6400 joint_samples 18 [978942, 484096]
processed_samples 6401 unjoint_samples 6400 joint_samples 17 [1046849, 594416]
processed_samples 6401 unjoint_samples 6400 joint_samples 17 [1046849, 594416]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a1567a40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1043298, 674887]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
processed_samples 6500 unjoint_samples 6500 joint_samples 19 [1029273, 525405]
processed_samples 6500 unjoint_samples 6500 joint_samples 19 [821475, 1037049]
processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1036519, 628310]
processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1043298, 674887]
processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1046833, 674978]
processed_samples 6500 unjoint_samples 6500 joint_samples 18 [978942, 711322]
processed_samples 6500 unjoint_samples 6500 joint_samples 19 [1029273, 525405]
processed_samples 6500 unjoint_samples 6500 joint_samples 19 [821475, 1037049]
processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1036519, 628310]
processed_samples 6500 unjoint_samples 6500 joint_samples 19 [554838, 1028150]
processed_samples 6500 unjoint_samples 6500 joint_samples 19 [554838, 1028150]
processed_samples 6500 unjoint_samples 6500 joint_samples 18 [1046833, 674978]
processed_samples 6501 unjoint_samples 6500 joint_samples 17 [1046849, 926259]
processed_samples 6500 unjoint_samples 6500 joint_samples 18 [978942, 711322]
processed_samples 6501 unjoint_samples 6500 joint_samples 17 [1046849, 926259]
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x5592a0a47f40] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0a47f40] mmco: unref short failure
[h264 @ 0x5592a0a47f40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0a47f40] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 6600 unjoint_samples 6600 joint_samples 20 [1018273, 134172]
processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1046833, 976085]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1043298, 1010321]
processed_samples 6600 unjoint_samples 6600 joint_samples 19 [1029273, 815528]
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1038824, 1028947]
processed_samples 6600 unjoint_samples 6600 joint_samples 19 [797667, 1028150]
processed_samples 6601 unjoint_samples 6600 joint_samples 18 [1046849, 415225]
processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1036519, 917332]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 6600 unjoint_samples 6600 joint_samples 19 [1029273, 815528]
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
processed_samples 6600 unjoint_samples 6600 joint_samples 20 [1018273, 134172]
processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1043298, 1010321]
processed_samples 6600 unjoint_samples 6600 joint_samples 19 [797667, 1028150]
processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1036519, 917332]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1046833, 976085]
processed_samples 6601 unjoint_samples 6600 joint_samples 18 [1046849, 415225]
processed_samples 6600 unjoint_samples 6600 joint_samples 18 [1038824, 1028947]
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0a47f40] mmco: unref short failure
[h264 @ 0x5592a0a47f40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x5592a0a47f40] mmco: unref short failure
[h264 @ 0x5592a0a47f40] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
processed_samples 6700 unjoint_samples 6700 joint_samples 20 [137178, 1032587]
processed_samples 6700 unjoint_samples 6700 joint_samples 20 [137053, 1045577]
processed_samples 6700 unjoint_samples 6700 joint_samples 20 [137178, 1032587]
processed_samples 6700 unjoint_samples 6700 joint_samples 19 [174838, 1046330]
processed_samples 6700 unjoint_samples 6700 joint_samples 20 [137053, 1045577]
processed_samples 6700 unjoint_samples 6700 joint_samples 19 [290293, 1039495]
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
processed_samples 6700 unjoint_samples 6700 joint_samples 20 [1018273, 390792]
processed_samples 6700 unjoint_samples 6700 joint_samples 19 [174838, 1046330]
processed_samples 6700 unjoint_samples 6700 joint_samples 19 [290293, 1039495]
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
processed_samples 6700 unjoint_samples 6700 joint_samples 20 [1018273, 390792]
processed_samples 6700 unjoint_samples 6700 joint_samples 19 [224992, 1046229]
processed_samples 6700 unjoint_samples 6700 joint_samples 19 [224992, 1046229]
processed_samples 6700 unjoint_samples 6700 joint_samples 19 [1043624, 330680]
processed_samples 6701 unjoint_samples 6700 joint_samples 18 [1046849, 737474]
processed_samples 6700 unjoint_samples 6700 joint_samples 19 [1043624, 330680]
processed_samples 6701 unjoint_samples 6700 joint_samples 18 [1046849, 737474]
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
processed_samples 6800 unjoint_samples 6800 joint_samples 19 [466638, 1046330]
processed_samples 6800 unjoint_samples 6800 joint_samples 19 [598196, 1046229]
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
processed_samples 6800 unjoint_samples 6800 joint_samples 20 [1018273, 833568]
processed_samples 6800 unjoint_samples 6800 joint_samples 19 [466638, 1046330]
processed_samples 6800 unjoint_samples 6800 joint_samples 20 [485242, 1032587]
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
processed_samples 6800 unjoint_samples 6800 joint_samples 19 [1043624, 602791]
processed_samples 6801 unjoint_samples 6800 joint_samples 19 [191691, 967655]
processed_samples 6800 unjoint_samples 6800 joint_samples 19 [598196, 1046229]
processed_samples 6800 unjoint_samples 6800 joint_samples 19 [538888, 1039495]
processed_samples 6800 unjoint_samples 6800 joint_samples 20 [475890, 1045577]
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
processed_samples 6800 unjoint_samples 6800 joint_samples 20 [1018273, 833568]
processed_samples 6800 unjoint_samples 6800 joint_samples 20 [485242, 1032587]
processed_samples 6800 unjoint_samples 6800 joint_samples 19 [1043624, 602791]
processed_samples 6801 unjoint_samples 6800 joint_samples 19 [191691, 967655]
processed_samples 6800 unjoint_samples 6800 joint_samples 19 [538888, 1039495]
processed_samples 6800 unjoint_samples 6800 joint_samples 20 [475890, 1045577]
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x55d97a55ba40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 6900 unjoint_samples 6900 joint_samples 21 [19959, 1046730]
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 6900 unjoint_samples 6900 joint_samples 21 [19959, 1046730]
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
processed_samples 6900 unjoint_samples 6900 joint_samples 20 [768357, 1032587]
processed_samples 6900 unjoint_samples 6900 joint_samples 20 [768357, 1032587]
processed_samples 6900 unjoint_samples 6900 joint_samples 20 [860510, 1045577]
processed_samples 6900 unjoint_samples 6900 joint_samples 19 [933617, 1046229]
processed_samples 6900 unjoint_samples 6900 joint_samples 19 [781606, 1039495]
processed_samples 6900 unjoint_samples 6900 joint_samples 19 [781606, 1039495]
processed_samples 6900 unjoint_samples 6900 joint_samples 19 [933617, 1046229]
processed_samples 6901 unjoint_samples 6900 joint_samples 19 [589434, 967655]
processed_samples 6901 unjoint_samples 6900 joint_samples 19 [589434, 967655]
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
processed_samples 6900 unjoint_samples 6900 joint_samples 20 [860510, 1045577]
processed_samples 6900 unjoint_samples 6900 joint_samples 19 [1043624, 947409]
processed_samples 6900 unjoint_samples 6900 joint_samples 19 [778078, 1046330]
processed_samples 6900 unjoint_samples 6900 joint_samples 19 [778078, 1046330]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
processed_samples 6900 unjoint_samples 6900 joint_samples 19 [1043624, 947409]
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [96511, 1041913]
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [96511, 1041913]
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1046987, 116515]
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1026676, 134992]
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1026676, 134992]
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [1046987, 116515]
processed_samples 7000 unjoint_samples 7000 joint_samples 21 [383845, 1046730]
processed_samples 7000 unjoint_samples 7000 joint_samples 21 [383845, 1046730]
processed_samples 7000 unjoint_samples 7000 joint_samples 21 [21138, 1047925]
processed_samples 7000 unjoint_samples 7000 joint_samples 21 [21138, 1047925]
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [158002, 1037014]
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [158002, 1037014]
processed_samples 7001 unjoint_samples 7000 joint_samples 19 [947415, 967655]
processed_samples 7001 unjoint_samples 7000 joint_samples 19 [947415, 967655]
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [996115, 1032587]
processed_samples 7000 unjoint_samples 7000 joint_samples 20 [996115, 1032587]
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97a820dc0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
processed_samples 7100 unjoint_samples 7100 joint_samples 21 [633091, 1046730]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
processed_samples 7100 unjoint_samples 7100 joint_samples 21 [633091, 1046730]
processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1026676, 445144]
processed_samples 7100 unjoint_samples 7100 joint_samples 20 [385944, 1041913]
processed_samples 7100 unjoint_samples 7100 joint_samples 20 [385944, 1041913]
processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1026676, 445144]
processed_samples 7100 unjoint_samples 7100 joint_samples 21 [226860, 1038877]
processed_samples 7100 unjoint_samples 7100 joint_samples 21 [382178, 1047925]
processed_samples 7100 unjoint_samples 7100 joint_samples 21 [382178, 1047925]
processed_samples 7100 unjoint_samples 7100 joint_samples 21 [226860, 1038877]
processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1046987, 362869]
processed_samples 7100 unjoint_samples 7100 joint_samples 20 [1046987, 362869]
processed_samples 7101 unjoint_samples 7100 joint_samples 20 [1006726, 246065]
processed_samples 7101 unjoint_samples 7100 joint_samples 20 [1006726, 246065]
processed_samples 7100 unjoint_samples 7100 joint_samples 20 [419270, 1037014]
processed_samples 7100 unjoint_samples 7100 joint_samples 20 [419270, 1037014]
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1046987, 687301]
processed_samples 7200 unjoint_samples 7200 joint_samples 21 [631974, 1047925]
processed_samples 7200 unjoint_samples 7200 joint_samples 21 [465010, 1038877]
processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1046987, 687301]
processed_samples 7200 unjoint_samples 7200 joint_samples 21 [631974, 1047925]
processed_samples 7200 unjoint_samples 7200 joint_samples 21 [966805, 1046730]
processed_samples 7200 unjoint_samples 7200 joint_samples 21 [465010, 1038877]
processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1026676, 737242]
processed_samples 7200 unjoint_samples 7200 joint_samples 20 [1026676, 737242]
processed_samples 7201 unjoint_samples 7200 joint_samples 20 [1006726, 606824]
processed_samples 7200 unjoint_samples 7200 joint_samples 20 [741076, 1041913]
processed_samples 7200 unjoint_samples 7200 joint_samples 21 [966805, 1046730]
processed_samples 7200 unjoint_samples 7200 joint_samples 20 [741076, 1041913]
processed_samples 7200 unjoint_samples 7200 joint_samples 20 [719581, 1037014]
processed_samples 7201 unjoint_samples 7200 joint_samples 20 [1006726, 606824]
processed_samples 7200 unjoint_samples 7200 joint_samples 20 [719581, 1037014]
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
processed_samples 7300 unjoint_samples 7300 joint_samples 22 [1045593, 170813]
processed_samples 7300 unjoint_samples 7300 joint_samples 21 [62159, 1045915]
processed_samples 7300 unjoint_samples 7300 joint_samples 21 [944589, 1047925]
processed_samples 7300 unjoint_samples 7300 joint_samples 20 [984984, 1041913]
processed_samples 7300 unjoint_samples 7300 joint_samples 20 [1047399, 1048127]
processed_samples 7301 unjoint_samples 7300 joint_samples 20 [1006726, 911396]
processed_samples 7300 unjoint_samples 7300 joint_samples 21 [703883, 1038877]
processed_samples 7300 unjoint_samples 7300 joint_samples 22 [1045593, 170813]
processed_samples 7300 unjoint_samples 7300 joint_samples 21 [62159, 1045915]
processed_samples 7300 unjoint_samples 7300 joint_samples 21 [944589, 1047925]
processed_samples 7300 unjoint_samples 7300 joint_samples 20 [998699, 1037014]
processed_samples 7300 unjoint_samples 7300 joint_samples 20 [1047399, 1048127]
processed_samples 7300 unjoint_samples 7300 joint_samples 20 [984984, 1041913]
processed_samples 7301 unjoint_samples 7300 joint_samples 20 [1006726, 911396]
processed_samples 7300 unjoint_samples 7300 joint_samples 21 [703883, 1038877]
processed_samples 7300 unjoint_samples 7300 joint_samples 20 [998699, 1037014]
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [377307, 1045915]
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [377307, 1045915]
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1047399, 255899]
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1047399, 255899]
processed_samples 7400 unjoint_samples 7400 joint_samples 22 [1045593, 442560]
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1033555, 289982]
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1033555, 289982]
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [266161, 1041913]
processed_samples 7400 unjoint_samples 7400 joint_samples 22 [1045593, 442560]
processed_samples 7401 unjoint_samples 7400 joint_samples 21 [1041293, 131202]
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [266161, 1041913]
processed_samples 7401 unjoint_samples 7400 joint_samples 21 [1041293, 131202]
processed_samples 7400 unjoint_samples 7400 joint_samples 22 [186931, 1047925]
processed_samples 7400 unjoint_samples 7400 joint_samples 22 [186931, 1047925]
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1021773, 1038877]
processed_samples 7400 unjoint_samples 7400 joint_samples 21 [1021773, 1038877]
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
processed_samples 7500 unjoint_samples 7500 joint_samples 21 [1033555, 760022]
processed_samples 7500 unjoint_samples 7500 joint_samples 22 [203691, 1047977]
processed_samples 7500 unjoint_samples 7500 joint_samples 22 [526501, 1047925]
processed_samples 7500 unjoint_samples 7500 joint_samples 21 [1047399, 552778]
processed_samples 7500 unjoint_samples 7500 joint_samples 21 [602413, 1041913]
processed_samples 7500 unjoint_samples 7500 joint_samples 22 [1045593, 758343]
processed_samples 7501 unjoint_samples 7500 joint_samples 21 [1041293, 415816]
processed_samples 7500 unjoint_samples 7500 joint_samples 21 [846865, 1045915]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
processed_samples 7500 unjoint_samples 7500 joint_samples 21 [1033555, 760022]
processed_samples 7500 unjoint_samples 7500 joint_samples 22 [203691, 1047977]
processed_samples 7500 unjoint_samples 7500 joint_samples 22 [526501, 1047925]
processed_samples 7500 unjoint_samples 7500 joint_samples 21 [1047399, 552778]
processed_samples 7500 unjoint_samples 7500 joint_samples 21 [846865, 1045915]
processed_samples 7500 unjoint_samples 7500 joint_samples 21 [602413, 1041913]
processed_samples 7501 unjoint_samples 7500 joint_samples 21 [1041293, 415816]
processed_samples 7500 unjoint_samples 7500 joint_samples 22 [1045593, 758343]
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 7600 unjoint_samples 7600 joint_samples 21 [1033555, 998637]
processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1003477, 217495]
processed_samples 7600 unjoint_samples 7600 joint_samples 22 [656550, 1047977]
processed_samples 7600 unjoint_samples 7600 joint_samples 21 [1047399, 845573]
processed_samples 7600 unjoint_samples 7600 joint_samples 22 [809617, 1047925]
processed_samples 7601 unjoint_samples 7600 joint_samples 21 [1041293, 710270]
processed_samples 7600 unjoint_samples 7600 joint_samples 21 [890116, 1041913]
processed_samples 7600 unjoint_samples 7600 joint_samples 21 [1047399, 845573]
processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1045593, 998345]
processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1045593, 998345]
processed_samples 7600 unjoint_samples 7600 joint_samples 21 [1033555, 998637]
processed_samples 7600 unjoint_samples 7600 joint_samples 22 [1003477, 217495]
processed_samples 7600 unjoint_samples 7600 joint_samples 22 [809617, 1047925]
processed_samples 7600 unjoint_samples 7600 joint_samples 22 [656550, 1047977]
processed_samples 7600 unjoint_samples 7600 joint_samples 21 [890116, 1041913]
processed_samples 7601 unjoint_samples 7600 joint_samples 21 [1041293, 710270]
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
processed_samples 7700 unjoint_samples 7700 joint_samples 23 [1026218, 47675]
processed_samples 7700 unjoint_samples 7700 joint_samples 23 [1026218, 47675]
processed_samples 7700 unjoint_samples 7700 joint_samples 22 [130309, 1046448]
processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1047399, 100363]
processed_samples 7700 unjoint_samples 7700 joint_samples 22 [130309, 1046448]
processed_samples 7700 unjoint_samples 7700 joint_samples 23 [1047235, 61520]
processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1047399, 100363]
processed_samples 7700 unjoint_samples 7700 joint_samples 23 [1047235, 61520]
processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1003477, 438264]
processed_samples 7700 unjoint_samples 7700 joint_samples 22 [1003477, 438264]
processed_samples 7700 unjoint_samples 7700 joint_samples 23 [1045593, 238913]
processed_samples 7700 unjoint_samples 7700 joint_samples 23 [1045593, 238913]
processed_samples 7700 unjoint_samples 7700 joint_samples 22 [205229, 1047523]
processed_samples 7700 unjoint_samples 7700 joint_samples 22 [205229, 1047523]
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 7701 unjoint_samples 7700 joint_samples 21 [1041293, 1030419]
processed_samples 7701 unjoint_samples 7700 joint_samples 21 [1041293, 1030419]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
processed_samples 7800 unjoint_samples 7800 joint_samples 22 [474782, 1046448]
processed_samples 7800 unjoint_samples 7800 joint_samples 23 [1047235, 326067]
processed_samples 7800 unjoint_samples 7800 joint_samples 23 [1045593, 616592]
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
processed_samples 7800 unjoint_samples 7800 joint_samples 23 [1026218, 313058]
processed_samples 7800 unjoint_samples 7800 joint_samples 22 [633896, 1047523]
processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1047399, 441349]
processed_samples 7801 unjoint_samples 7800 joint_samples 22 [335474, 1037240]
processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1003477, 785102]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
processed_samples 7800 unjoint_samples 7800 joint_samples 23 [1047235, 326067]
processed_samples 7800 unjoint_samples 7800 joint_samples 23 [1026218, 313058]
processed_samples 7800 unjoint_samples 7800 joint_samples 22 [474782, 1046448]
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
processed_samples 7800 unjoint_samples 7800 joint_samples 23 [1045593, 616592]
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1047399, 441349]
processed_samples 7800 unjoint_samples 7800 joint_samples 22 [633896, 1047523]
processed_samples 7801 unjoint_samples 7800 joint_samples 22 [335474, 1037240]
processed_samples 7800 unjoint_samples 7800 joint_samples 22 [1003477, 785102]
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
processed_samples 7900 unjoint_samples 7900 joint_samples 22 [754090, 1046448]
processed_samples 7900 unjoint_samples 7900 joint_samples 22 [754090, 1046448]
processed_samples 7900 unjoint_samples 7900 joint_samples 23 [151609, 1026157]
processed_samples 7900 unjoint_samples 7900 joint_samples 23 [151609, 1026157]
processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1026218, 628787]
processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1047235, 589319]
processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1026218, 628787]
processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1047235, 589319]
processed_samples 7900 unjoint_samples 7900 joint_samples 22 [1047399, 722225]
processed_samples 7900 unjoint_samples 7900 joint_samples 22 [1047399, 722225]
processed_samples 7900 unjoint_samples 7900 joint_samples 22 [897707, 1047523]
processed_samples 7900 unjoint_samples 7900 joint_samples 22 [897707, 1047523]
processed_samples 7901 unjoint_samples 7900 joint_samples 22 [713026, 1037240]
processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1045593, 882906]
processed_samples 7900 unjoint_samples 7900 joint_samples 23 [1045593, 882906]
processed_samples 7901 unjoint_samples 7900 joint_samples 22 [713026, 1037240]
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[mov,mp4,m4a,3gp,3g2,mj2 @ 0x55d98337c9c0] stream 0, offset 0x90050f: partial file
[mov,mp4,m4a,3gp,3g2,mj2 @ 0x5592a4e016c0] stream 0, offset 0x90050f: partial file
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
processed_samples 8000 unjoint_samples 8000 joint_samples 24 [198879, 1045379]
processed_samples 8000 unjoint_samples 8000 joint_samples 23 [451245, 1026157]
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
processed_samples 8000 unjoint_samples 8000 joint_samples 22 [1047399, 1031799]
processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1026218, 902937]
processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1023768, 29483]
processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1046622, 189652]
processed_samples 8001 unjoint_samples 8000 joint_samples 22 [1010088, 1037240]
processed_samples 8000 unjoint_samples 8000 joint_samples 24 [198879, 1045379]
processed_samples 8000 unjoint_samples 8000 joint_samples 23 [451245, 1026157]
processed_samples 8000 unjoint_samples 8000 joint_samples 22 [1047399, 1031799]
processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1023768, 29483]
processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1046622, 189652]
processed_samples 8001 unjoint_samples 8000 joint_samples 23 [1047235, 918376]
processed_samples 8001 unjoint_samples 8000 joint_samples 22 [1010088, 1037240]
processed_samples 8000 unjoint_samples 8000 joint_samples 23 [1026218, 902937]
processed_samples 8001 unjoint_samples 8000 joint_samples 23 [1047235, 918376]
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1047399, 236906]
processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1023768, 405945]
processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1023768, 405945]
processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1047399, 236906]
processed_samples 8100 unjoint_samples 8100 joint_samples 23 [796172, 1026157]
processed_samples 8100 unjoint_samples 8100 joint_samples 24 [549830, 1045379]
processed_samples 8100 unjoint_samples 8100 joint_samples 23 [796172, 1026157]
processed_samples 8100 unjoint_samples 8100 joint_samples 24 [549830, 1045379]
processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1046622, 624549]
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
processed_samples 8101 unjoint_samples 8100 joint_samples 23 [293298, 1046235]
processed_samples 8101 unjoint_samples 8100 joint_samples 23 [293298, 1046235]
processed_samples 8100 unjoint_samples 8100 joint_samples 24 [86198, 1047079]
processed_samples 8100 unjoint_samples 8100 joint_samples 24 [86198, 1047079]
processed_samples 8100 unjoint_samples 8100 joint_samples 23 [1046622, 624549]
processed_samples 8101 unjoint_samples 8100 joint_samples 24 [257200, 1046385]
processed_samples 8101 unjoint_samples 8100 joint_samples 24 [257200, 1046385]
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 8200 unjoint_samples 8200 joint_samples 23 [1023768, 740667]
processed_samples 8200 unjoint_samples 8200 joint_samples 24 [346781, 1047079]
processed_samples 8200 unjoint_samples 8200 joint_samples 23 [1047399, 511717]
processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1024865, 29721]
processed_samples 8200 unjoint_samples 8200 joint_samples 23 [1046622, 846580]
processed_samples 8200 unjoint_samples 8200 joint_samples 24 [876629, 1045379]
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
processed_samples 8200 unjoint_samples 8200 joint_samples 23 [1023768, 740667]
processed_samples 8200 unjoint_samples 8200 joint_samples 24 [346781, 1047079]
processed_samples 8200 unjoint_samples 8200 joint_samples 23 [1047399, 511717]
processed_samples 8200 unjoint_samples 8200 joint_samples 24 [876629, 1045379]
processed_samples 8200 unjoint_samples 8200 joint_samples 24 [1024865, 29721]
processed_samples 8200 unjoint_samples 8200 joint_samples 23 [1046622, 846580]
processed_samples 8201 unjoint_samples 8200 joint_samples 24 [542172, 1046385]
processed_samples 8201 unjoint_samples 8200 joint_samples 23 [624260, 1046235]
processed_samples 8201 unjoint_samples 8200 joint_samples 24 [542172, 1046385]
processed_samples 8201 unjoint_samples 8200 joint_samples 23 [624260, 1046235]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
processed_samples 8300 unjoint_samples 8300 joint_samples 25 [1023835, 190881]
processed_samples 8300 unjoint_samples 8300 joint_samples 25 [1023835, 190881]
processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1046622, 70108]
processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1024865, 341076]
processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1046622, 70108]
processed_samples 8300 unjoint_samples 8300 joint_samples 24 [1024865, 341076]
processed_samples 8300 unjoint_samples 8300 joint_samples 23 [1047399, 802233]
processed_samples 8300 unjoint_samples 8300 joint_samples 24 [665127, 1047079]
processed_samples 8300 unjoint_samples 8300 joint_samples 23 [1047399, 802233]
processed_samples 8300 unjoint_samples 8300 joint_samples 23 [1029216, 1031096]
processed_samples 8300 unjoint_samples 8300 joint_samples 23 [1029216, 1031096]
processed_samples 8300 unjoint_samples 8300 joint_samples 24 [665127, 1047079]
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
processed_samples 8301 unjoint_samples 8300 joint_samples 23 [922193, 1046235]
processed_samples 8301 unjoint_samples 8300 joint_samples 24 [884694, 1046385]
processed_samples 8301 unjoint_samples 8300 joint_samples 23 [922193, 1046235]
processed_samples 8301 unjoint_samples 8300 joint_samples 24 [884694, 1046385]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 8400 unjoint_samples 8400 joint_samples 25 [1023835, 508234]
processed_samples 8400 unjoint_samples 8400 joint_samples 25 [1023835, 508234]
processed_samples 8400 unjoint_samples 8400 joint_samples 24 [926996, 1047079]
processed_samples 8400 unjoint_samples 8400 joint_samples 24 [395202, 1046387]
processed_samples 8400 unjoint_samples 8400 joint_samples 24 [926996, 1047079]
processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1046622, 360885]
processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1024865, 642655]
processed_samples 8400 unjoint_samples 8400 joint_samples 24 [395202, 1046387]
processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1046622, 360885]
processed_samples 8401 unjoint_samples 8400 joint_samples 24 [278047, 1046235]
processed_samples 8401 unjoint_samples 8400 joint_samples 25 [1034524, 157409]
processed_samples 8400 unjoint_samples 8400 joint_samples 24 [1024865, 642655]
processed_samples 8400 unjoint_samples 8400 joint_samples 23 [1047399, 1018121]
processed_samples 8401 unjoint_samples 8400 joint_samples 24 [278047, 1046235]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 8401 unjoint_samples 8400 joint_samples 25 [1034524, 157409]
processed_samples 8400 unjoint_samples 8400 joint_samples 23 [1047399, 1018121]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
processed_samples 8500 unjoint_samples 8500 joint_samples 24 [314623, 1039223]
processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1026501, 141210]
processed_samples 8500 unjoint_samples 8500 joint_samples 24 [790494, 1046387]
processed_samples 8500 unjoint_samples 8500 joint_samples 24 [1046622, 632385]
processed_samples 8501 unjoint_samples 8500 joint_samples 25 [1034524, 424489]
processed_samples 8500 unjoint_samples 8500 joint_samples 24 [1024865, 1017562]
processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1023835, 907186]
processed_samples 8501 unjoint_samples 8500 joint_samples 24 [619946, 1046235]
processed_samples 8500 unjoint_samples 8500 joint_samples 24 [314623, 1039223]
processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1026501, 141210]
processed_samples 8500 unjoint_samples 8500 joint_samples 24 [790494, 1046387]
processed_samples 8500 unjoint_samples 8500 joint_samples 25 [1023835, 907186]
processed_samples 8500 unjoint_samples 8500 joint_samples 24 [1046622, 632385]
processed_samples 8501 unjoint_samples 8500 joint_samples 25 [1034524, 424489]
processed_samples 8500 unjoint_samples 8500 joint_samples 24 [1024865, 1017562]
processed_samples 8501 unjoint_samples 8500 joint_samples 24 [619946, 1046235]
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a97d90c0] mmco: unref short failure
[h264 @ 0x5592a97d90c0] mmco: unref short failure
[h264 @ 0x5592a97d90c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1046545, 274166]
processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1046545, 274166]
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
processed_samples 8600 unjoint_samples 8600 joint_samples 26 [1043809, 154162]
processed_samples 8600 unjoint_samples 8600 joint_samples 26 [1043809, 154162]
processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1026501, 457008]
processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1026501, 457008]
processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1047511, 3169]
processed_samples 8600 unjoint_samples 8600 joint_samples 24 [655883, 1039223]
processed_samples 8600 unjoint_samples 8600 joint_samples 24 [655883, 1039223]
processed_samples 8600 unjoint_samples 8600 joint_samples 25 [1047511, 3169]
processed_samples 8600 unjoint_samples 8600 joint_samples 24 [1046622, 911940]
processed_samples 8600 unjoint_samples 8600 joint_samples 24 [1046622, 911940]
processed_samples 8601 unjoint_samples 8600 joint_samples 24 [933275, 1046235]
processed_samples 8601 unjoint_samples 8600 joint_samples 24 [933275, 1046235]
processed_samples 8601 unjoint_samples 8600 joint_samples 25 [1034524, 829729]
processed_samples 8601 unjoint_samples 8600 joint_samples 25 [1034524, 829729]
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a1345a80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a1345a80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97b06f840] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97b06f840] mmco: unref short failure
[h264 @ 0x55d97b06f840] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1047511, 286733]
processed_samples 8700 unjoint_samples 8700 joint_samples 26 [1043809, 465548]
processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1047511, 286733]
processed_samples 8700 unjoint_samples 8700 joint_samples 26 [1043809, 465548]
processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1046695, 140766]
processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1046695, 140766]
processed_samples 8700 unjoint_samples 8700 joint_samples 24 [952733, 1039223]
processed_samples 8701 unjoint_samples 8700 joint_samples 26 [1037554, 136078]
processed_samples 8701 unjoint_samples 8700 joint_samples 26 [1037554, 136078]
processed_samples 8701 unjoint_samples 8700 joint_samples 25 [1045726, 248142]
processed_samples 8700 unjoint_samples 8700 joint_samples 24 [952733, 1039223]
processed_samples 8701 unjoint_samples 8700 joint_samples 25 [1045726, 248142]
processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1026501, 692903]
processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1026501, 692903]
processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1046545, 669054]
processed_samples 8700 unjoint_samples 8700 joint_samples 25 [1046545, 669054]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97b06f840] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1047846, 103006]
processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1043809, 850737]
[h264 @ 0x55d97f274580] mmco: unref short failure
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1047511, 591453]
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1046695, 380853]
processed_samples 8801 unjoint_samples 8800 joint_samples 26 [1037554, 553759]
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1028454, 1027760]
[h264 @ 0x5592a529c8c0] mmco: unref short failure
processed_samples 8801 unjoint_samples 8800 joint_samples 25 [1045726, 719489]
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1046545, 924431]
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
processed_samples 8800 unjoint_samples 8800 joint_samples 26 [1043809, 850737]
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1047846, 103006]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1047511, 591453]
processed_samples 8801 unjoint_samples 8800 joint_samples 26 [1037554, 553759]
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1046695, 380853]
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1028454, 1027760]
processed_samples 8801 unjoint_samples 8800 joint_samples 25 [1045726, 719489]
processed_samples 8800 unjoint_samples 8800 joint_samples 25 [1046545, 924431]
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a4a37940] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
processed_samples 8900 unjoint_samples 8900 joint_samples 27 [114044, 1046579]
processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1046545, 214895]
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
processed_samples 8900 unjoint_samples 8900 joint_samples 25 [1047846, 370080]
processed_samples 8900 unjoint_samples 8900 joint_samples 25 [1047511, 910166]
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1043715, 304636]
processed_samples 8900 unjoint_samples 8900 joint_samples 27 [114044, 1046579]
processed_samples 8900 unjoint_samples 8900 joint_samples 25 [1046695, 698381]
processed_samples 8900 unjoint_samples 8900 joint_samples 25 [1047846, 370080]
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1046545, 214895]
processed_samples 8901 unjoint_samples 8900 joint_samples 26 [1037554, 841788]
processed_samples 8901 unjoint_samples 8900 joint_samples 25 [1045726, 1038158]
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
processed_samples 8900 unjoint_samples 8900 joint_samples 25 [1046695, 698381]
processed_samples 8900 unjoint_samples 8900 joint_samples 25 [1047511, 910166]
processed_samples 8901 unjoint_samples 8900 joint_samples 26 [1037554, 841788]
processed_samples 8900 unjoint_samples 8900 joint_samples 26 [1043715, 304636]
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a10b4380] mmco: unref short failure
processed_samples 8901 unjoint_samples 8900 joint_samples 25 [1045726, 1038158]
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x5592a4ae3700] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
processed_samples 9000 unjoint_samples 9000 joint_samples 26 [215997, 1015787]
processed_samples 9000 unjoint_samples 9000 joint_samples 26 [215997, 1015787]
processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1046695, 31683]
processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1046695, 31683]
processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1043715, 674205]
processed_samples 9000 unjoint_samples 9000 joint_samples 27 [407230, 1046579]
processed_samples 9001 unjoint_samples 9000 joint_samples 27 [1047850, 19970]
processed_samples 9000 unjoint_samples 9000 joint_samples 27 [407230, 1046579]
processed_samples 9001 unjoint_samples 9000 joint_samples 27 [1047850, 19970]
processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1046545, 516269]
processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1046545, 516269]
processed_samples 9000 unjoint_samples 9000 joint_samples 26 [1043715, 674205]
processed_samples 9001 unjoint_samples 9000 joint_samples 26 [1045726, 293856]
processed_samples 9001 unjoint_samples 9000 joint_samples 26 [1045726, 293856]
processed_samples 9000 unjoint_samples 9000 joint_samples 25 [1047846, 820674]
processed_samples 9000 unjoint_samples 9000 joint_samples 25 [1047846, 820674]
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a13f7a00] mmco: unref short failure
[h264 @ 0x5592a13f7a00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1046545, 1041295]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [127662, 1047542]
processed_samples 9100 unjoint_samples 9100 joint_samples 27 [725131, 1046579]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [624628, 1015787]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1046695, 404401]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [127662, 1047542]
processed_samples 9100 unjoint_samples 9100 joint_samples 27 [725131, 1046579]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1043715, 968668]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [624628, 1015787]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1046545, 1041295]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1046695, 404401]
processed_samples 9100 unjoint_samples 9100 joint_samples 26 [1043715, 968668]
processed_samples 9101 unjoint_samples 9100 joint_samples 27 [1047850, 300085]
processed_samples 9101 unjoint_samples 9100 joint_samples 26 [1045726, 489938]
processed_samples 9101 unjoint_samples 9100 joint_samples 26 [1045726, 489938]
processed_samples 9101 unjoint_samples 9100 joint_samples 27 [1047850, 300085]
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
processed_samples 9200 unjoint_samples 9200 joint_samples 26 [903499, 1015787]
processed_samples 9200 unjoint_samples 9200 joint_samples 26 [456112, 1047542]
processed_samples 9200 unjoint_samples 9200 joint_samples 27 [214964, 1047233]
processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1048055, 135483]
processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1031983, 1046579]
processed_samples 9201 unjoint_samples 9200 joint_samples 27 [1047850, 510924]
processed_samples 9200 unjoint_samples 9200 joint_samples 26 [1046695, 733649]
processed_samples 9201 unjoint_samples 9200 joint_samples 26 [1045726, 744380]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
processed_samples 9200 unjoint_samples 9200 joint_samples 26 [903499, 1015787]
processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1048055, 135483]
processed_samples 9200 unjoint_samples 9200 joint_samples 27 [214964, 1047233]
processed_samples 9200 unjoint_samples 9200 joint_samples 26 [456112, 1047542]
processed_samples 9201 unjoint_samples 9200 joint_samples 27 [1047850, 510924]
processed_samples 9200 unjoint_samples 9200 joint_samples 26 [1046695, 733649]
processed_samples 9200 unjoint_samples 9200 joint_samples 27 [1031983, 1046579]
processed_samples 9201 unjoint_samples 9200 joint_samples 26 [1045726, 744380]
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a5201c80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
processed_samples 9300 unjoint_samples 9300 joint_samples 28 [215580, 1046579]
processed_samples 9300 unjoint_samples 9300 joint_samples 28 [215580, 1046579]
processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1046695, 79101]
processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1048055, 378654]
processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1046695, 79101]
processed_samples 9300 unjoint_samples 9300 joint_samples 27 [1048055, 378654]
processed_samples 9300 unjoint_samples 9300 joint_samples 27 [256587, 1045171]
processed_samples 9300 unjoint_samples 9300 joint_samples 27 [256587, 1045171]
processed_samples 9301 unjoint_samples 9300 joint_samples 27 [1045726, 22027]
processed_samples 9301 unjoint_samples 9300 joint_samples 27 [1045726, 22027]
processed_samples 9300 unjoint_samples 9300 joint_samples 26 [728707, 1047542]
processed_samples 9300 unjoint_samples 9300 joint_samples 27 [637360, 1047233]
processed_samples 9300 unjoint_samples 9300 joint_samples 27 [637360, 1047233]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 9300 unjoint_samples 9300 joint_samples 26 [728707, 1047542]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 9301 unjoint_samples 9300 joint_samples 27 [1047850, 854377]
[h264 @ 0x5592a159cec0] mmco: unref short failure
processed_samples 9301 unjoint_samples 9300 joint_samples 27 [1047850, 854377]
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
processed_samples 9400 unjoint_samples 9400 joint_samples 26 [999971, 1047542]
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 9400 unjoint_samples 9400 joint_samples 28 [529546, 1046579]
[h264 @ 0x5592a31fac00] mmco: unref short failure
processed_samples 9400 unjoint_samples 9400 joint_samples 27 [574251, 1045171]
processed_samples 9401 unjoint_samples 9400 joint_samples 28 [132695, 1040900]
processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1046695, 442678]
processed_samples 9401 unjoint_samples 9400 joint_samples 27 [1045726, 272230]
processed_samples 9400 unjoint_samples 9400 joint_samples 27 [962553, 1047233]
processed_samples 9400 unjoint_samples 9400 joint_samples 26 [999971, 1047542]
processed_samples 9400 unjoint_samples 9400 joint_samples 28 [529546, 1046579]
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a0fef300] mmco: unref short failure
processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1048055, 724287]
processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1046695, 442678]
processed_samples 9400 unjoint_samples 9400 joint_samples 27 [574251, 1045171]
processed_samples 9401 unjoint_samples 9400 joint_samples 28 [132695, 1040900]
processed_samples 9401 unjoint_samples 9400 joint_samples 27 [1045726, 272230]
processed_samples 9400 unjoint_samples 9400 joint_samples 27 [1048055, 724287]
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 9400 unjoint_samples 9400 joint_samples 27 [962553, 1047233]
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a0fef300] mmco: unref short failure
[h264 @ 0x5592a0fef300] mmco: unref short failure
[h264 @ 0x5592a0fef300] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
processed_samples 9500 unjoint_samples 9500 joint_samples 28 [866112, 1046579]
processed_samples 9500 unjoint_samples 9500 joint_samples 28 [866112, 1046579]
processed_samples 9500 unjoint_samples 9500 joint_samples 28 [1048055, 13701]
processed_samples 9500 unjoint_samples 9500 joint_samples 28 [1048055, 13701]
processed_samples 9500 unjoint_samples 9500 joint_samples 27 [1044889, 264086]
processed_samples 9500 unjoint_samples 9500 joint_samples 27 [1044889, 264086]
processed_samples 9500 unjoint_samples 9500 joint_samples 28 [243912, 1047233]
processed_samples 9500 unjoint_samples 9500 joint_samples 28 [243912, 1047233]
processed_samples 9501 unjoint_samples 9500 joint_samples 28 [350830, 1040900]
processed_samples 9501 unjoint_samples 9500 joint_samples 28 [350830, 1040900]
processed_samples 9500 unjoint_samples 9500 joint_samples 27 [1046695, 713267]
processed_samples 9500 unjoint_samples 9500 joint_samples 27 [869725, 1045171]
processed_samples 9500 unjoint_samples 9500 joint_samples 27 [869725, 1045171]
processed_samples 9500 unjoint_samples 9500 joint_samples 27 [1046695, 713267]
processed_samples 9501 unjoint_samples 9500 joint_samples 27 [1045726, 577126]
processed_samples 9501 unjoint_samples 9500 joint_samples 27 [1045726, 577126]
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
processed_samples 9600 unjoint_samples 9600 joint_samples 29 [1027325, 181253]
processed_samples 9600 unjoint_samples 9600 joint_samples 28 [112748, 1047737]
processed_samples 9600 unjoint_samples 9600 joint_samples 29 [1027325, 181253]
processed_samples 9600 unjoint_samples 9600 joint_samples 27 [1044889, 573838]
[h264 @ 0x55d97facaf00] mmco: unref short failure
processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1048055, 310761]
processed_samples 9600 unjoint_samples 9600 joint_samples 27 [1046695, 1020421]
processed_samples 9600 unjoint_samples 9600 joint_samples 28 [572145, 1047233]
processed_samples 9600 unjoint_samples 9600 joint_samples 28 [1048055, 310761]
[h264 @ 0x5592a618b9c0] mmco: unref short failure
processed_samples 9600 unjoint_samples 9600 joint_samples 27 [1046695, 1020421]
processed_samples 9600 unjoint_samples 9600 joint_samples 28 [572145, 1047233]
processed_samples 9600 unjoint_samples 9600 joint_samples 28 [112748, 1047737]
processed_samples 9601 unjoint_samples 9600 joint_samples 28 [729999, 1040900]
processed_samples 9600 unjoint_samples 9600 joint_samples 27 [1044889, 573838]
processed_samples 9601 unjoint_samples 9600 joint_samples 28 [729999, 1040900]
processed_samples 9601 unjoint_samples 9600 joint_samples 27 [1045726, 951370]
processed_samples 9601 unjoint_samples 9600 joint_samples 27 [1045726, 951370]
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0fef300] mmco: unref short failure
[h264 @ 0x5592a0fef300] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
processed_samples 9700 unjoint_samples 9700 joint_samples 29 [1027325, 539558]
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 9700 unjoint_samples 9700 joint_samples 28 [1046695, 235819]
processed_samples 9700 unjoint_samples 9700 joint_samples 28 [1046695, 235819]
processed_samples 9700 unjoint_samples 9700 joint_samples 28 [1048055, 568414]
processed_samples 9700 unjoint_samples 9700 joint_samples 28 [464661, 1047737]
processed_samples 9700 unjoint_samples 9700 joint_samples 29 [1027325, 539558]
processed_samples 9700 unjoint_samples 9700 joint_samples 28 [464661, 1047737]
processed_samples 9700 unjoint_samples 9700 joint_samples 28 [854352, 1047233]
processed_samples 9700 unjoint_samples 9700 joint_samples 28 [854352, 1047233]
processed_samples 9701 unjoint_samples 9700 joint_samples 28 [241069, 1044753]
processed_samples 9701 unjoint_samples 9700 joint_samples 28 [241069, 1044753]
processed_samples 9700 unjoint_samples 9700 joint_samples 28 [1048055, 568414]
processed_samples 9700 unjoint_samples 9700 joint_samples 27 [1044889, 994963]
processed_samples 9701 unjoint_samples 9700 joint_samples 28 [990166, 1040900]
processed_samples 9700 unjoint_samples 9700 joint_samples 27 [1044889, 994963]
processed_samples 9701 unjoint_samples 9700 joint_samples 28 [990166, 1040900]
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 9800 unjoint_samples 9800 joint_samples 29 [94013, 1047762]
processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1047539, 219592]
processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1027325, 797949]
processed_samples 9800 unjoint_samples 9800 joint_samples 28 [711704, 1047737]
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1046695, 555580]
processed_samples 9800 unjoint_samples 9800 joint_samples 29 [94013, 1047762]
processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1047539, 219592]
[h264 @ 0x55d9826bc180] mmco: unref short failure
processed_samples 9800 unjoint_samples 9800 joint_samples 28 [711704, 1047737]
processed_samples 9800 unjoint_samples 9800 joint_samples 29 [1027325, 797949]
processed_samples 9801 unjoint_samples 9800 joint_samples 28 [537630, 1044753]
processed_samples 9801 unjoint_samples 9800 joint_samples 29 [258937, 1045118]
processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1046695, 555580]
processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1048055, 943897]
[h264 @ 0x5592a529c8c0] mmco: unref short failure
processed_samples 9801 unjoint_samples 9800 joint_samples 28 [537630, 1044753]
processed_samples 9801 unjoint_samples 9800 joint_samples 29 [258937, 1045118]
processed_samples 9800 unjoint_samples 9800 joint_samples 28 [1048055, 943897]
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1048055, 175025]
processed_samples 9900 unjoint_samples 9900 joint_samples 29 [393564, 1047762]
processed_samples 9900 unjoint_samples 9900 joint_samples 28 [1047539, 471827]
processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1047907, 83039]
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
processed_samples 9900 unjoint_samples 9900 joint_samples 28 [1046695, 875842]
processed_samples 9900 unjoint_samples 9900 joint_samples 30 [1035887, 36622]
processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1048055, 175025]
processed_samples 9900 unjoint_samples 9900 joint_samples 29 [393564, 1047762]
processed_samples 9900 unjoint_samples 9900 joint_samples 28 [1047539, 471827]
processed_samples 9900 unjoint_samples 9900 joint_samples 29 [1047907, 83039]
processed_samples 9900 unjoint_samples 9900 joint_samples 28 [1046695, 875842]
processed_samples 9900 unjoint_samples 9900 joint_samples 30 [1035887, 36622]
processed_samples 9901 unjoint_samples 9900 joint_samples 28 [982206, 1044753]
processed_samples 9901 unjoint_samples 9900 joint_samples 28 [982206, 1044753]
processed_samples 9901 unjoint_samples 9900 joint_samples 29 [565176, 1045118]
processed_samples 9901 unjoint_samples 9900 joint_samples 29 [565176, 1045118]
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a1a6f700] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
processed_samples 10000 unjoint_samples 10000 joint_samples 29 [629045, 1047762]
processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1047907, 425841]
processed_samples 10000 unjoint_samples 10000 joint_samples 30 [1035887, 279868]
processed_samples 10000 unjoint_samples 10000 joint_samples 29 [96036, 1044188]
processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1048055, 466886]
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
processed_samples 10000 unjoint_samples 10000 joint_samples 28 [1047539, 786272]
processed_samples 10000 unjoint_samples 10000 joint_samples 29 [629045, 1047762]
processed_samples 10001 unjoint_samples 10000 joint_samples 29 [841070, 1045118]
processed_samples 10001 unjoint_samples 10000 joint_samples 29 [247113, 1047477]
processed_samples 10000 unjoint_samples 10000 joint_samples 30 [1035887, 279868]
[h264 @ 0x55d97b679340] mmco: unref short failure
processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1047907, 425841]
processed_samples 10000 unjoint_samples 10000 joint_samples 29 [1048055, 466886]
processed_samples 10000 unjoint_samples 10000 joint_samples 29 [96036, 1044188]
[h264 @ 0x5592a23df380] mmco: unref short failure
processed_samples 10000 unjoint_samples 10000 joint_samples 28 [1047539, 786272]
[h264 @ 0x55d97b543c80] mmco: unref short failure
processed_samples 10001 unjoint_samples 10000 joint_samples 29 [247113, 1047477]
[h264 @ 0x5592a48aaa80] mmco: unref short failure
processed_samples 10001 unjoint_samples 10000 joint_samples 29 [841070, 1045118]
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [195648, 1046559]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [195648, 1046559]
processed_samples 10100 unjoint_samples 10100 joint_samples 30 [1035887, 684902]
processed_samples 10100 unjoint_samples 10100 joint_samples 30 [1035887, 684902]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [453711, 1044188]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [453711, 1044188]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [1048055, 712506]
processed_samples 10101 unjoint_samples 10100 joint_samples 30 [29494, 1048149]
processed_samples 10101 unjoint_samples 10100 joint_samples 30 [29494, 1048149]
processed_samples 10101 unjoint_samples 10100 joint_samples 29 [757687, 1047477]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [983918, 1047762]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [1048055, 712506]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [983918, 1047762]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [1047907, 906666]
processed_samples 10100 unjoint_samples 10100 joint_samples 29 [1047907, 906666]
processed_samples 10101 unjoint_samples 10100 joint_samples 29 [757687, 1047477]
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a1a6f700] mmco: unref short failure
[h264 @ 0x5592a1a6f700] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1017748, 328115]
processed_samples 10200 unjoint_samples 10200 joint_samples 29 [770293, 1044188]
processed_samples 10200 unjoint_samples 10200 joint_samples 29 [578826, 1046559]
[h264 @ 0x5592a1a6da40] mmco: unref short failure
processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1035887, 973572]
processed_samples 10200 unjoint_samples 10200 joint_samples 29 [1048055, 988692]
processed_samples 10201 unjoint_samples 10200 joint_samples 30 [325024, 1048149]
processed_samples 10200 unjoint_samples 10200 joint_samples 30 [168449, 1040234]
processed_samples 10201 unjoint_samples 10200 joint_samples 29 [1036848, 1047477]
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1017748, 328115]
processed_samples 10200 unjoint_samples 10200 joint_samples 29 [578826, 1046559]
processed_samples 10200 unjoint_samples 10200 joint_samples 29 [770293, 1044188]
[h264 @ 0x55d97e08af80] mmco: unref short failure
processed_samples 10200 unjoint_samples 10200 joint_samples 30 [1035887, 973572]
processed_samples 10200 unjoint_samples 10200 joint_samples 30 [168449, 1040234]
processed_samples 10200 unjoint_samples 10200 joint_samples 29 [1048055, 988692]
processed_samples 10201 unjoint_samples 10200 joint_samples 30 [325024, 1048149]
processed_samples 10201 unjoint_samples 10200 joint_samples 29 [1036848, 1047477]
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
processed_samples 10300 unjoint_samples 10300 joint_samples 30 [214291, 1046916]
processed_samples 10300 unjoint_samples 10300 joint_samples 31 [1046195, 220481]
processed_samples 10300 unjoint_samples 10300 joint_samples 30 [559669, 1040234]
processed_samples 10300 unjoint_samples 10300 joint_samples 30 [214291, 1046916]
processed_samples 10300 unjoint_samples 10300 joint_samples 31 [1046195, 220481]
[h264 @ 0x55d97adfba00] mmco: unref short failure
processed_samples 10300 unjoint_samples 10300 joint_samples 30 [559669, 1040234]
processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1017748, 606779]
[h264 @ 0x5592a47dee40] mmco: unref short failure
processed_samples 10300 unjoint_samples 10300 joint_samples 30 [1017748, 606779]
processed_samples 10300 unjoint_samples 10300 joint_samples 29 [1004675, 1046559]
[h264 @ 0x55d97e08af80] mmco: unref short failure
processed_samples 10300 unjoint_samples 10300 joint_samples 29 [1004675, 1046559]
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 10300 unjoint_samples 10300 joint_samples 30 [9788, 1046481]
processed_samples 10300 unjoint_samples 10300 joint_samples 30 [9788, 1046481]
processed_samples 10301 unjoint_samples 10300 joint_samples 30 [334831, 1047477]
processed_samples 10301 unjoint_samples 10300 joint_samples 30 [334831, 1047477]
processed_samples 10301 unjoint_samples 10300 joint_samples 30 [625426, 1048149]
processed_samples 10301 unjoint_samples 10300 joint_samples 30 [625426, 1048149]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b6f1f80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
processed_samples 10400 unjoint_samples 10400 joint_samples 31 [1046195, 602221]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [881877, 1040234]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1032310, 358690]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [411829, 1046916]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1017748, 936055]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [233008, 1046481]
processed_samples 10401 unjoint_samples 10400 joint_samples 30 [667959, 1047477]
processed_samples 10401 unjoint_samples 10400 joint_samples 30 [958514, 1048149]
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [881877, 1040234]
processed_samples 10400 unjoint_samples 10400 joint_samples 31 [1046195, 602221]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1032310, 358690]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [411829, 1046916]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [1017748, 936055]
processed_samples 10400 unjoint_samples 10400 joint_samples 30 [233008, 1046481]
processed_samples 10401 unjoint_samples 10400 joint_samples 30 [667959, 1047477]
processed_samples 10401 unjoint_samples 10400 joint_samples 30 [958514, 1048149]
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 10500 unjoint_samples 10500 joint_samples 31 [159167, 1046273]
processed_samples 10500 unjoint_samples 10500 joint_samples 31 [159167, 1046273]
processed_samples 10500 unjoint_samples 10500 joint_samples 31 [379247, 1046802]
processed_samples 10500 unjoint_samples 10500 joint_samples 31 [379247, 1046802]
processed_samples 10500 unjoint_samples 10500 joint_samples 30 [1032310, 726185]
processed_samples 10500 unjoint_samples 10500 joint_samples 30 [1032310, 726185]
processed_samples 10500 unjoint_samples 10500 joint_samples 30 [713548, 1046916]
processed_samples 10500 unjoint_samples 10500 joint_samples 30 [713548, 1046916]
processed_samples 10501 unjoint_samples 10500 joint_samples 31 [12564, 1047477]
processed_samples 10501 unjoint_samples 10500 joint_samples 31 [12564, 1047477]
processed_samples 10501 unjoint_samples 10500 joint_samples 31 [1047896, 152801]
processed_samples 10501 unjoint_samples 10500 joint_samples 31 [1047896, 152801]
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
processed_samples 10500 unjoint_samples 10500 joint_samples 30 [538421, 1046481]
processed_samples 10500 unjoint_samples 10500 joint_samples 30 [538421, 1046481]
processed_samples 10500 unjoint_samples 10500 joint_samples 31 [1046195, 878730]
processed_samples 10500 unjoint_samples 10500 joint_samples 31 [1046195, 878730]
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
processed_samples 10600 unjoint_samples 10600 joint_samples 32 [206632, 1043212]
processed_samples 10600 unjoint_samples 10600 joint_samples 30 [970271, 1046916]
processed_samples 10601 unjoint_samples 10600 joint_samples 31 [364095, 1047477]
processed_samples 10600 unjoint_samples 10600 joint_samples 31 [842413, 1046802]
processed_samples 10600 unjoint_samples 10600 joint_samples 31 [538631, 1046273]
processed_samples 10601 unjoint_samples 10600 joint_samples 31 [1047896, 560162]
processed_samples 10600 unjoint_samples 10600 joint_samples 30 [863740, 1046481]
processed_samples 10600 unjoint_samples 10600 joint_samples 30 [1032310, 988382]
processed_samples 10600 unjoint_samples 10600 joint_samples 32 [206632, 1043212]
processed_samples 10600 unjoint_samples 10600 joint_samples 31 [538631, 1046273]
processed_samples 10600 unjoint_samples 10600 joint_samples 31 [842413, 1046802]
processed_samples 10600 unjoint_samples 10600 joint_samples 30 [1032310, 988382]
processed_samples 10600 unjoint_samples 10600 joint_samples 30 [970271, 1046916]
processed_samples 10601 unjoint_samples 10600 joint_samples 31 [364095, 1047477]
processed_samples 10601 unjoint_samples 10600 joint_samples 31 [1047896, 560162]
processed_samples 10600 unjoint_samples 10600 joint_samples 30 [863740, 1046481]
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
processed_samples 10700 unjoint_samples 10700 joint_samples 31 [346452, 1043770]
processed_samples 10700 unjoint_samples 10700 joint_samples 31 [346452, 1043770]
processed_samples 10700 unjoint_samples 10700 joint_samples 31 [1046395, 172139]
processed_samples 10700 unjoint_samples 10700 joint_samples 31 [1046395, 172139]
processed_samples 10700 unjoint_samples 10700 joint_samples 32 [497554, 1043212]
processed_samples 10700 unjoint_samples 10700 joint_samples 32 [277906, 1046802]
processed_samples 10700 unjoint_samples 10700 joint_samples 32 [277906, 1046802]
processed_samples 10700 unjoint_samples 10700 joint_samples 32 [497554, 1043212]
processed_samples 10700 unjoint_samples 10700 joint_samples 31 [925454, 1046273]
processed_samples 10700 unjoint_samples 10700 joint_samples 31 [171564, 1046481]
processed_samples 10700 unjoint_samples 10700 joint_samples 31 [171564, 1046481]
processed_samples 10700 unjoint_samples 10700 joint_samples 31 [925454, 1046273]
processed_samples 10701 unjoint_samples 10700 joint_samples 31 [678407, 1047477]
processed_samples 10701 unjoint_samples 10700 joint_samples 31 [678407, 1047477]
processed_samples 10701 unjoint_samples 10700 joint_samples 31 [1047896, 811926]
processed_samples 10701 unjoint_samples 10700 joint_samples 31 [1047896, 811926]
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97a9c6040] mmco: unref short failure
[h264 @ 0x55d97a9c6040] mmco: unref short failure
[h264 @ 0x55d97a9c6040] mmco: unref short failure
[h264 @ 0x55d97a9c6040] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
processed_samples 10800 unjoint_samples 10800 joint_samples 32 [147407, 1046273]
processed_samples 10800 unjoint_samples 10800 joint_samples 31 [483485, 1046481]
processed_samples 10800 unjoint_samples 10800 joint_samples 31 [1046395, 486950]
processed_samples 10800 unjoint_samples 10800 joint_samples 31 [681039, 1043770]
processed_samples 10800 unjoint_samples 10800 joint_samples 32 [689752, 1046802]
[h264 @ 0x5592a48560c0] mmco: unref short failure
processed_samples 10800 unjoint_samples 10800 joint_samples 32 [750334, 1043212]
processed_samples 10801 unjoint_samples 10800 joint_samples 31 [1003295, 1047477]
processed_samples 10800 unjoint_samples 10800 joint_samples 32 [147407, 1046273]
processed_samples 10801 unjoint_samples 10800 joint_samples 32 [165673, 1031423]
processed_samples 10800 unjoint_samples 10800 joint_samples 31 [681039, 1043770]
processed_samples 10800 unjoint_samples 10800 joint_samples 32 [689752, 1046802]
processed_samples 10800 unjoint_samples 10800 joint_samples 31 [1046395, 486950]
processed_samples 10800 unjoint_samples 10800 joint_samples 31 [483485, 1046481]
[h264 @ 0x55d97ac78600] mmco: unref short failure
processed_samples 10800 unjoint_samples 10800 joint_samples 32 [750334, 1043212]
processed_samples 10801 unjoint_samples 10800 joint_samples 32 [165673, 1031423]
processed_samples 10801 unjoint_samples 10800 joint_samples 31 [1003295, 1047477]
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97facaf00] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97facaf00] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
processed_samples 10900 unjoint_samples 10900 joint_samples 33 [1028498, 82228]
processed_samples 10900 unjoint_samples 10900 joint_samples 33 [1028498, 82228]
processed_samples 10900 unjoint_samples 10900 joint_samples 31 [796343, 1046481]
processed_samples 10900 unjoint_samples 10900 joint_samples 31 [796343, 1046481]
processed_samples 10900 unjoint_samples 10900 joint_samples 32 [1044700, 13624]
processed_samples 10900 unjoint_samples 10900 joint_samples 32 [1044700, 13624]
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
processed_samples 10901 unjoint_samples 10900 joint_samples 32 [454296, 1031423]
processed_samples 10901 unjoint_samples 10900 joint_samples 32 [454296, 1031423]
processed_samples 10901 unjoint_samples 10900 joint_samples 32 [1046790, 221736]
processed_samples 10901 unjoint_samples 10900 joint_samples 32 [1046790, 221736]
processed_samples 10900 unjoint_samples 10900 joint_samples 32 [989160, 1046802]
processed_samples 10900 unjoint_samples 10900 joint_samples 32 [404588, 1046273]
processed_samples 10900 unjoint_samples 10900 joint_samples 32 [404588, 1046273]
processed_samples 10900 unjoint_samples 10900 joint_samples 31 [1046395, 840342]
processed_samples 10900 unjoint_samples 10900 joint_samples 32 [989160, 1046802]
processed_samples 10900 unjoint_samples 10900 joint_samples 31 [1046395, 840342]
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a13cdd80] mmco: unref short failure
[h264 @ 0x5592a13cdd80] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a48d1c00] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a10512c0] mmco: unref short failure
[h264 @ 0x5592a10512c0] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 11000 unjoint_samples 11000 joint_samples 32 [1044700, 352888]
processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1028498, 360469]
processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1044784, 196762]
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
processed_samples 11000 unjoint_samples 11000 joint_samples 32 [1038072, 67630]
processed_samples 11000 unjoint_samples 11000 joint_samples 32 [732184, 1046273]
processed_samples 11000 unjoint_samples 11000 joint_samples 31 [1046395, 1035881]
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1028498, 360469]
processed_samples 11001 unjoint_samples 11000 joint_samples 32 [1046790, 509554]
processed_samples 11000 unjoint_samples 11000 joint_samples 32 [1044700, 352888]
processed_samples 11000 unjoint_samples 11000 joint_samples 33 [1044784, 196762]
processed_samples 11001 unjoint_samples 11000 joint_samples 32 [771386, 1031423]
processed_samples 11000 unjoint_samples 11000 joint_samples 32 [732184, 1046273]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
processed_samples 11000 unjoint_samples 11000 joint_samples 32 [1038072, 67630]
processed_samples 11000 unjoint_samples 11000 joint_samples 31 [1046395, 1035881]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
processed_samples 11001 unjoint_samples 11000 joint_samples 32 [1046790, 509554]
processed_samples 11001 unjoint_samples 11000 joint_samples 32 [771386, 1031423]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1028498, 689411]
processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1044784, 448219]
processed_samples 11100 unjoint_samples 11100 joint_samples 32 [1038072, 463855]
processed_samples 11100 unjoint_samples 11100 joint_samples 32 [1044700, 641124]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 11100 unjoint_samples 11100 joint_samples 32 [342135, 1042587]
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 11101 unjoint_samples 11100 joint_samples 32 [1046790, 1000470]
processed_samples 11101 unjoint_samples 11100 joint_samples 32 [1042477, 1041808]
processed_samples 11100 unjoint_samples 11100 joint_samples 32 [959066, 1046273]
processed_samples 11100 unjoint_samples 11100 joint_samples 32 [342135, 1042587]
processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1028498, 689411]
processed_samples 11100 unjoint_samples 11100 joint_samples 33 [1044784, 448219]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 11100 unjoint_samples 11100 joint_samples 32 [1044700, 641124]
processed_samples 11100 unjoint_samples 11100 joint_samples 32 [1038072, 463855]
processed_samples 11100 unjoint_samples 11100 joint_samples 32 [959066, 1046273]
[h264 @ 0x55d97d921300] mmco: unref short failure
processed_samples 11101 unjoint_samples 11100 joint_samples 32 [1046790, 1000470]
processed_samples 11101 unjoint_samples 11100 joint_samples 32 [1042477, 1041808]
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1025973, 250158]
processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1028498, 903838]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1044784, 736709]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 11200 unjoint_samples 11200 joint_samples 32 [1044700, 937900]
processed_samples 11200 unjoint_samples 11200 joint_samples 32 [692775, 1042587]
processed_samples 11201 unjoint_samples 11200 joint_samples 33 [1046826, 276241]
processed_samples 11201 unjoint_samples 11200 joint_samples 33 [1047114, 520025]
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
processed_samples 11200 unjoint_samples 11200 joint_samples 32 [1038072, 810053]
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1028498, 903838]
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1025973, 250158]
processed_samples 11200 unjoint_samples 11200 joint_samples 33 [1044784, 736709]
processed_samples 11200 unjoint_samples 11200 joint_samples 32 [692775, 1042587]
processed_samples 11201 unjoint_samples 11200 joint_samples 33 [1046826, 276241]
processed_samples 11201 unjoint_samples 11200 joint_samples 33 [1047114, 520025]
processed_samples 11200 unjoint_samples 11200 joint_samples 32 [1044700, 937900]
processed_samples 11200 unjoint_samples 11200 joint_samples 32 [1038072, 810053]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
processed_samples 11300 unjoint_samples 11300 joint_samples 34 [1046606, 172401]
processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1046726, 171922]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1025973, 647086]
processed_samples 11300 unjoint_samples 11300 joint_samples 32 [990760, 1042587]
processed_samples 11300 unjoint_samples 11300 joint_samples 33 [347190, 986161]
processed_samples 11301 unjoint_samples 11300 joint_samples 33 [1046826, 567467]
processed_samples 11301 unjoint_samples 11300 joint_samples 33 [1047114, 862028]
processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1044784, 1011526]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 11300 unjoint_samples 11300 joint_samples 34 [1046606, 172401]
processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1046726, 171922]
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1025973, 647086]
processed_samples 11300 unjoint_samples 11300 joint_samples 33 [347190, 986161]
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
processed_samples 11300 unjoint_samples 11300 joint_samples 32 [990760, 1042587]
processed_samples 11301 unjoint_samples 11300 joint_samples 33 [1046826, 567467]
processed_samples 11301 unjoint_samples 11300 joint_samples 33 [1047114, 862028]
processed_samples 11300 unjoint_samples 11300 joint_samples 33 [1044784, 1011526]
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a0d56b40] mmco: unref short failure
[h264 @ 0x5592a0d56b40] mmco: unref short failure
[h264 @ 0x5592a0d56b40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1046726, 455886]
processed_samples 11400 unjoint_samples 11400 joint_samples 34 [317854, 1040034]
processed_samples 11400 unjoint_samples 11400 joint_samples 34 [1046606, 401631]
processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1044351, 300268]
processed_samples 11400 unjoint_samples 11400 joint_samples 33 [690964, 986161]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
processed_samples 11401 unjoint_samples 11400 joint_samples 34 [87736, 1044856]
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1040389, 1039000]
[h264 @ 0x5592a9a59e00] mmco: unref short failure
processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1046726, 455886]
processed_samples 11400 unjoint_samples 11400 joint_samples 34 [317854, 1040034]
processed_samples 11400 unjoint_samples 11400 joint_samples 34 [1046606, 401631]
processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1044351, 300268]
processed_samples 11401 unjoint_samples 11400 joint_samples 33 [1046826, 948987]
processed_samples 11400 unjoint_samples 11400 joint_samples 33 [690964, 986161]
processed_samples 11401 unjoint_samples 11400 joint_samples 34 [87736, 1044856]
processed_samples 11400 unjoint_samples 11400 joint_samples 33 [1040389, 1039000]
processed_samples 11401 unjoint_samples 11400 joint_samples 33 [1046826, 948987]
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 11500 unjoint_samples 11500 joint_samples 34 [661727, 1040034]
processed_samples 11500 unjoint_samples 11500 joint_samples 34 [1046606, 791519]
processed_samples 11501 unjoint_samples 11500 joint_samples 34 [269906, 1040767]
processed_samples 11500 unjoint_samples 11500 joint_samples 34 [661727, 1040034]
processed_samples 11500 unjoint_samples 11500 joint_samples 33 [1044351, 609157]
processed_samples 11500 unjoint_samples 11500 joint_samples 34 [1040389, 296215]
processed_samples 11500 unjoint_samples 11500 joint_samples 33 [969795, 986161]
processed_samples 11501 unjoint_samples 11500 joint_samples 34 [469381, 1044856]
processed_samples 11500 unjoint_samples 11500 joint_samples 34 [1046606, 791519]
processed_samples 11500 unjoint_samples 11500 joint_samples 34 [1040389, 296215]
processed_samples 11500 unjoint_samples 11500 joint_samples 33 [1046726, 729256]
processed_samples 11501 unjoint_samples 11500 joint_samples 34 [269906, 1040767]
processed_samples 11500 unjoint_samples 11500 joint_samples 33 [1044351, 609157]
processed_samples 11501 unjoint_samples 11500 joint_samples 34 [469381, 1044856]
processed_samples 11500 unjoint_samples 11500 joint_samples 33 [969795, 986161]
processed_samples 11500 unjoint_samples 11500 joint_samples 33 [1046726, 729256]
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
local_disk/cognitron_vl/cognitron_vl/data/processor/image_processor.py:265: ResourceWarning: unclosed file <_io.BufferedReader name='datasets/LMM/lmms-lab/LLaVA-OneVision-Data/ureader_qa/ureader-instruction-1.0/TextVQA/train_images/ca6e0d79c7cc20bd.jpg'>
  image = Image.open(img_or_path).convert("RGB")
local_disk/cognitron_vl/cognitron_vl/data/processor/image_processor.py:265: ResourceWarning: unclosed file <_io.BufferedReader name='datasets/LMM/lmms-lab/LLaVA-OneVision-Data/ureader_qa/ureader-instruction-1.0/TextVQA/train_images/ca6e0d79c7cc20bd.jpg'>
  image = Image.open(img_or_path).convert("RGB")
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
processed_samples 11600 unjoint_samples 11600 joint_samples 35 [59887, 1035292]
processed_samples 11600 unjoint_samples 11600 joint_samples 35 [59887, 1035292]
processed_samples 11600 unjoint_samples 11600 joint_samples 34 [1040389, 608098]
processed_samples 11600 unjoint_samples 11600 joint_samples 33 [1046726, 954600]
processed_samples 11600 unjoint_samples 11600 joint_samples 33 [1044351, 912983]
processed_samples 11600 unjoint_samples 11600 joint_samples 34 [961791, 1040034]
processed_samples 11600 unjoint_samples 11600 joint_samples 34 [351276, 988161]
processed_samples 11600 unjoint_samples 11600 joint_samples 34 [1040389, 608098]
processed_samples 11601 unjoint_samples 11600 joint_samples 34 [817757, 1044856]
processed_samples 11600 unjoint_samples 11600 joint_samples 34 [351276, 988161]
processed_samples 11600 unjoint_samples 11600 joint_samples 33 [1044351, 912983]
processed_samples 11601 unjoint_samples 11600 joint_samples 34 [716242, 1040767]
processed_samples 11601 unjoint_samples 11600 joint_samples 34 [817757, 1044856]
processed_samples 11600 unjoint_samples 11600 joint_samples 34 [961791, 1040034]
processed_samples 11600 unjoint_samples 11600 joint_samples 33 [1046726, 954600]
processed_samples 11601 unjoint_samples 11600 joint_samples 34 [716242, 1040767]
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x55d97cda95c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 11700 unjoint_samples 11700 joint_samples 35 [323777, 1035292]
processed_samples 11700 unjoint_samples 11700 joint_samples 35 [323777, 1035292]
processed_samples 11700 unjoint_samples 11700 joint_samples 35 [1046712, 185034]
[h264 @ 0x5592a10b5440] mmco: unref short failure
[h264 @ 0x5592a10b5440] mmco: unref short failure
processed_samples 11700 unjoint_samples 11700 joint_samples 35 [1046712, 185034]
processed_samples 11700 unjoint_samples 11700 joint_samples 34 [136794, 1046260]
processed_samples 11700 unjoint_samples 11700 joint_samples 34 [136794, 1046260]
processed_samples 11700 unjoint_samples 11700 joint_samples 34 [620874, 988161]
processed_samples 11700 unjoint_samples 11700 joint_samples 34 [174366, 1047981]
processed_samples 11700 unjoint_samples 11700 joint_samples 34 [620874, 988161]
processed_samples 11700 unjoint_samples 11700 joint_samples 34 [174366, 1047981]
processed_samples 11700 unjoint_samples 11700 joint_samples 34 [1040389, 971216]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
processed_samples 11701 unjoint_samples 11700 joint_samples 35 [992712, 238439]
processed_samples 11701 unjoint_samples 11700 joint_samples 35 [992712, 238439]
processed_samples 11701 unjoint_samples 11700 joint_samples 34 [1022159, 1040767]
processed_samples 11701 unjoint_samples 11700 joint_samples 34 [1022159, 1040767]
processed_samples 11700 unjoint_samples 11700 joint_samples 34 [1040389, 971216]
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a10b5440] mmco: unref short failure
[h264 @ 0x5592a10b5440] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97eb84d00] mmco: unref short failure
[h264 @ 0x5592a10b5440] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
processed_samples 11800 unjoint_samples 11800 joint_samples 35 [1044988, 275836]
processed_samples 11800 unjoint_samples 11800 joint_samples 34 [504008, 1047981]
processed_samples 11800 unjoint_samples 11800 joint_samples 34 [430935, 1046260]
processed_samples 11800 unjoint_samples 11800 joint_samples 35 [675240, 1035292]
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 11800 unjoint_samples 11800 joint_samples 35 [1046712, 602165]
processed_samples 11801 unjoint_samples 11800 joint_samples 35 [1047499, 267314]
processed_samples 11801 unjoint_samples 11800 joint_samples 35 [992712, 497538]
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
processed_samples 11800 unjoint_samples 11800 joint_samples 34 [983621, 988161]
processed_samples 11800 unjoint_samples 11800 joint_samples 35 [1044988, 275836]
processed_samples 11800 unjoint_samples 11800 joint_samples 34 [430935, 1046260]
processed_samples 11800 unjoint_samples 11800 joint_samples 34 [504008, 1047981]
processed_samples 11800 unjoint_samples 11800 joint_samples 35 [675240, 1035292]
processed_samples 11800 unjoint_samples 11800 joint_samples 35 [1046712, 602165]
processed_samples 11801 unjoint_samples 11800 joint_samples 35 [1047499, 267314]
processed_samples 11801 unjoint_samples 11800 joint_samples 35 [992712, 497538]
processed_samples 11800 unjoint_samples 11800 joint_samples 34 [983621, 988161]
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a2b5cac0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1045190, 142824]
processed_samples 11900 unjoint_samples 11900 joint_samples 35 [992593, 1035292]
processed_samples 11900 unjoint_samples 11900 joint_samples 34 [668392, 1046260]
processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1044988, 625593]
processed_samples 11901 unjoint_samples 11900 joint_samples 35 [1047499, 706395]
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
processed_samples 11901 unjoint_samples 11900 joint_samples 35 [992712, 805593]
processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1046712, 929319]
processed_samples 11900 unjoint_samples 11900 joint_samples 34 [739473, 1047981]
processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1045190, 142824]
[h264 @ 0x5592a50d5e80] mmco: unref short failure
processed_samples 11900 unjoint_samples 11900 joint_samples 35 [992593, 1035292]
processed_samples 11900 unjoint_samples 11900 joint_samples 34 [668392, 1046260]
processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1044988, 625593]
processed_samples 11901 unjoint_samples 11900 joint_samples 35 [1047499, 706395]
processed_samples 11901 unjoint_samples 11900 joint_samples 35 [992712, 805593]
processed_samples 11900 unjoint_samples 11900 joint_samples 35 [1046712, 929319]
processed_samples 11900 unjoint_samples 11900 joint_samples 34 [739473, 1047981]
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
processed_samples 12000 unjoint_samples 12000 joint_samples 36 [1029845, 345869]
processed_samples 12000 unjoint_samples 12000 joint_samples 36 [244062, 1046374]
processed_samples 12000 unjoint_samples 12000 joint_samples 34 [1040319, 1047981]
processed_samples 12000 unjoint_samples 12000 joint_samples 36 [1029845, 345869]
processed_samples 12000 unjoint_samples 12000 joint_samples 35 [1044988, 927875]
processed_samples 12000 unjoint_samples 12000 joint_samples 34 [1040319, 1047981]
processed_samples 12000 unjoint_samples 12000 joint_samples 36 [244062, 1046374]
processed_samples 12000 unjoint_samples 12000 joint_samples 35 [1045190, 592152]
processed_samples 12000 unjoint_samples 12000 joint_samples 35 [1045190, 592152]
processed_samples 12000 unjoint_samples 12000 joint_samples 35 [1044988, 927875]
processed_samples 12000 unjoint_samples 12000 joint_samples 34 [964172, 1046260]
processed_samples 12001 unjoint_samples 12000 joint_samples 36 [16433, 1040824]
processed_samples 12001 unjoint_samples 12000 joint_samples 36 [16433, 1040824]
processed_samples 12000 unjoint_samples 12000 joint_samples 34 [964172, 1046260]
processed_samples 12001 unjoint_samples 12000 joint_samples 36 [1030769, 61950]
processed_samples 12001 unjoint_samples 12000 joint_samples 36 [1030769, 61950]
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97b646440] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
processed_samples 12100 unjoint_samples 12100 joint_samples 36 [190140, 1046332]
processed_samples 12101 unjoint_samples 12100 joint_samples 36 [386141, 1040824]
processed_samples 12100 unjoint_samples 12100 joint_samples 35 [1045246, 138471]
processed_samples 12100 unjoint_samples 12100 joint_samples 35 [327585, 1047981]
processed_samples 12100 unjoint_samples 12100 joint_samples 36 [190140, 1046332]
processed_samples 12101 unjoint_samples 12100 joint_samples 36 [386141, 1040824]
processed_samples 12100 unjoint_samples 12100 joint_samples 36 [501238, 1046374]
processed_samples 12100 unjoint_samples 12100 joint_samples 36 [1029845, 683528]
processed_samples 12100 unjoint_samples 12100 joint_samples 35 [1045246, 138471]
processed_samples 12101 unjoint_samples 12100 joint_samples 36 [1030769, 306408]
processed_samples 12100 unjoint_samples 12100 joint_samples 35 [327585, 1047981]
processed_samples 12100 unjoint_samples 12100 joint_samples 35 [1045190, 862080]
processed_samples 12100 unjoint_samples 12100 joint_samples 36 [501238, 1046374]
processed_samples 12100 unjoint_samples 12100 joint_samples 36 [1029845, 683528]
processed_samples 12101 unjoint_samples 12100 joint_samples 36 [1030769, 306408]
processed_samples 12100 unjoint_samples 12100 joint_samples 35 [1045190, 862080]
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97ea8dc80] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
processed_samples 12200 unjoint_samples 12200 joint_samples 35 [1045246, 550066]
processed_samples 12200 unjoint_samples 12200 joint_samples 35 [1045246, 550066]
processed_samples 12200 unjoint_samples 12200 joint_samples 36 [1029845, 998429]
processed_samples 12200 unjoint_samples 12200 joint_samples 36 [72626, 1046419]
processed_samples 12200 unjoint_samples 12200 joint_samples 36 [1029845, 998429]
processed_samples 12200 unjoint_samples 12200 joint_samples 36 [72626, 1046419]
processed_samples 12200 unjoint_samples 12200 joint_samples 36 [986952, 1046374]
processed_samples 12201 unjoint_samples 12200 joint_samples 36 [854116, 1040824]
processed_samples 12200 unjoint_samples 12200 joint_samples 36 [533808, 1046332]
processed_samples 12201 unjoint_samples 12200 joint_samples 36 [1030769, 593956]
processed_samples 12201 unjoint_samples 12200 joint_samples 36 [854116, 1040824]
processed_samples 12200 unjoint_samples 12200 joint_samples 36 [986952, 1046374]
processed_samples 12200 unjoint_samples 12200 joint_samples 36 [533808, 1046332]
processed_samples 12200 unjoint_samples 12200 joint_samples 35 [644737, 1047981]
processed_samples 12201 unjoint_samples 12200 joint_samples 36 [1030769, 593956]
processed_samples 12200 unjoint_samples 12200 joint_samples 35 [644737, 1047981]
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
processed_samples 12300 unjoint_samples 12300 joint_samples 37 [1015547, 73831]
processed_samples 12300 unjoint_samples 12300 joint_samples 35 [1045246, 867193]
processed_samples 12300 unjoint_samples 12300 joint_samples 37 [998680, 335518]
processed_samples 12300 unjoint_samples 12300 joint_samples 37 [295721, 1032683]
processed_samples 12300 unjoint_samples 12300 joint_samples 37 [295721, 1032683]
processed_samples 12300 unjoint_samples 12300 joint_samples 36 [377132, 1046419]
processed_samples 12300 unjoint_samples 12300 joint_samples 37 [1015547, 73831]
processed_samples 12301 unjoint_samples 12300 joint_samples 37 [1042511, 102771]
processed_samples 12300 unjoint_samples 12300 joint_samples 37 [998680, 335518]
processed_samples 12300 unjoint_samples 12300 joint_samples 36 [377132, 1046419]
processed_samples 12300 unjoint_samples 12300 joint_samples 35 [923160, 1047981]
processed_samples 12300 unjoint_samples 12300 joint_samples 35 [1045246, 867193]
processed_samples 12300 unjoint_samples 12300 joint_samples 35 [923160, 1047981]
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
processed_samples 12301 unjoint_samples 12300 joint_samples 37 [1042511, 102771]
processed_samples 12301 unjoint_samples 12300 joint_samples 36 [1030769, 937214]
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
processed_samples 12301 unjoint_samples 12300 joint_samples 36 [1030769, 937214]
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a1226240] mmco: unref short failure
[h264 @ 0x5592a1226240] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x55d97eb34a00] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
processed_samples 12400 unjoint_samples 12400 joint_samples 36 [269937, 1015212]
processed_samples 12400 unjoint_samples 12400 joint_samples 36 [269937, 1015212]
[h264 @ 0x5592a0992b40] mmco: unref short failure
processed_samples 12400 unjoint_samples 12400 joint_samples 36 [1047531, 345969]
processed_samples 12400 unjoint_samples 12400 joint_samples 37 [1015547, 345246]
[h264 @ 0x5592a0992b40] mmco: unref short failure
processed_samples 12400 unjoint_samples 12400 joint_samples 37 [998680, 698721]
processed_samples 12400 unjoint_samples 12400 joint_samples 37 [541843, 1032683]
processed_samples 12400 unjoint_samples 12400 joint_samples 36 [721329, 1046419]
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
processed_samples 12400 unjoint_samples 12400 joint_samples 36 [1047531, 345969]
processed_samples 12400 unjoint_samples 12400 joint_samples 37 [1015547, 345246]
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1042511, 412249]
processed_samples 12400 unjoint_samples 12400 joint_samples 37 [541843, 1032683]
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
processed_samples 12400 unjoint_samples 12400 joint_samples 37 [998680, 698721]
processed_samples 12400 unjoint_samples 12400 joint_samples 36 [721329, 1046419]
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1042511, 412249]
processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1037722, 288310]
processed_samples 12401 unjoint_samples 12400 joint_samples 37 [1037722, 288310]
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
processed_samples 12500 unjoint_samples 12500 joint_samples 37 [804126, 1032683]
processed_samples 12500 unjoint_samples 12500 joint_samples 37 [1014874, 1013855]
processed_samples 12500 unjoint_samples 12500 joint_samples 36 [616484, 1015212]
processed_samples 12500 unjoint_samples 12500 joint_samples 36 [1047531, 665583]
processed_samples 12500 unjoint_samples 12500 joint_samples 37 [1015547, 682298]
processed_samples 12500 unjoint_samples 12500 joint_samples 36 [616484, 1015212]
processed_samples 12500 unjoint_samples 12500 joint_samples 36 [1021172, 1046419]
processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1037722, 581132]
processed_samples 12500 unjoint_samples 12500 joint_samples 37 [804126, 1032683]
processed_samples 12500 unjoint_samples 12500 joint_samples 36 [1047531, 665583]
processed_samples 12500 unjoint_samples 12500 joint_samples 37 [1015547, 682298]
processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1042511, 845766]
processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1037722, 581132]
processed_samples 12500 unjoint_samples 12500 joint_samples 37 [1014874, 1013855]
processed_samples 12500 unjoint_samples 12500 joint_samples 36 [1021172, 1046419]
processed_samples 12501 unjoint_samples 12500 joint_samples 37 [1042511, 845766]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
processed_samples 12600 unjoint_samples 12600 joint_samples 38 [1047628, 84417]
processed_samples 12600 unjoint_samples 12600 joint_samples 38 [1046848, 280848]
[h264 @ 0x5592a12fb640] mmco: unref short failure
processed_samples 12600 unjoint_samples 12600 joint_samples 36 [943466, 1015212]
processed_samples 12601 unjoint_samples 12600 joint_samples 38 [82317, 1047714]
processed_samples 12600 unjoint_samples 12600 joint_samples 36 [1047531, 989328]
processed_samples 12600 unjoint_samples 12600 joint_samples 37 [264687, 1046419]
processed_samples 12600 unjoint_samples 12600 joint_samples 37 [1015547, 949071]
processed_samples 12600 unjoint_samples 12600 joint_samples 38 [1047628, 84417]
processed_samples 12600 unjoint_samples 12600 joint_samples 38 [1046848, 280848]
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
processed_samples 12600 unjoint_samples 12600 joint_samples 36 [943466, 1015212]
processed_samples 12601 unjoint_samples 12600 joint_samples 38 [82317, 1047714]
processed_samples 12600 unjoint_samples 12600 joint_samples 36 [1047531, 989328]
processed_samples 12600 unjoint_samples 12600 joint_samples 37 [264687, 1046419]
processed_samples 12600 unjoint_samples 12600 joint_samples 37 [1015547, 949071]
processed_samples 12601 unjoint_samples 12600 joint_samples 37 [1037722, 1013095]
processed_samples 12601 unjoint_samples 12600 joint_samples 37 [1037722, 1013095]
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e5e39c0] mmco: unref short failure
[h264 @ 0x55d97e5e39c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97e5e39c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
processed_samples 12700 unjoint_samples 12700 joint_samples 37 [1045318, 146884]
processed_samples 12700 unjoint_samples 12700 joint_samples 37 [1045318, 146884]
processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1047628, 435681]
processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1047628, 435681]
processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1033637, 248293]
processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1033637, 248293]
processed_samples 12700 unjoint_samples 12700 joint_samples 37 [590959, 1046419]
processed_samples 12700 unjoint_samples 12700 joint_samples 37 [590959, 1046419]
processed_samples 12701 unjoint_samples 12700 joint_samples 38 [386969, 1047714]
processed_samples 12701 unjoint_samples 12700 joint_samples 38 [386969, 1047714]
processed_samples 12700 unjoint_samples 12700 joint_samples 37 [341489, 1045355]
processed_samples 12700 unjoint_samples 12700 joint_samples 37 [341489, 1045355]
processed_samples 12701 unjoint_samples 12700 joint_samples 38 [372144, 1026224]
processed_samples 12701 unjoint_samples 12700 joint_samples 38 [372144, 1026224]
processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1046848, 547609]
processed_samples 12700 unjoint_samples 12700 joint_samples 38 [1046848, 547609]
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a1ea7b40] mmco: unref short failure
[h264 @ 0x5592a1ea7b40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 12800 unjoint_samples 12800 joint_samples 37 [624349, 1045355]
processed_samples 12800 unjoint_samples 12800 joint_samples 37 [1045318, 432663]
processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1033637, 622243]
processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1047628, 722475]
processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1046848, 802951]
processed_samples 12801 unjoint_samples 12800 joint_samples 38 [654807, 1026224]
processed_samples 12800 unjoint_samples 12800 joint_samples 37 [1045318, 432663]
processed_samples 12801 unjoint_samples 12800 joint_samples 38 [734617, 1047714]
processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1047628, 722475]
processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1033637, 622243]
processed_samples 12800 unjoint_samples 12800 joint_samples 37 [624349, 1045355]
processed_samples 12800 unjoint_samples 12800 joint_samples 37 [865710, 1046419]
processed_samples 12800 unjoint_samples 12800 joint_samples 38 [1046848, 802951]
processed_samples 12801 unjoint_samples 12800 joint_samples 38 [654807, 1026224]
processed_samples 12801 unjoint_samples 12800 joint_samples 38 [734617, 1047714]
processed_samples 12800 unjoint_samples 12800 joint_samples 37 [865710, 1046419]
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d97e545600] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a1ea7b40] mmco: unref short failure
[h264 @ 0x5592a1ea7b40] mmco: unref short failure
[h264 @ 0x5592a1ea7b40] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
processed_samples 12900 unjoint_samples 12900 joint_samples 39 [1046848, 111886]
processed_samples 12900 unjoint_samples 12900 joint_samples 39 [1046848, 111886]
processed_samples 12900 unjoint_samples 12900 joint_samples 37 [1045318, 720307]
processed_samples 12900 unjoint_samples 12900 joint_samples 37 [1045318, 720307]
processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1046619, 115142]
processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1046619, 115142]
processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1047628, 966059]
processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1033637, 894698]
processed_samples 12900 unjoint_samples 12900 joint_samples 37 [967056, 1045355]
processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1033637, 894698]
[h264 @ 0x55d97a5c0900] mmco: unref short failure
processed_samples 12900 unjoint_samples 12900 joint_samples 38 [1047628, 966059]
processed_samples 12900 unjoint_samples 12900 joint_samples 37 [967056, 1045355]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 12901 unjoint_samples 12900 joint_samples 38 [936498, 1047714]
processed_samples 12901 unjoint_samples 12900 joint_samples 38 [936498, 1047714]
processed_samples 12901 unjoint_samples 12900 joint_samples 38 [977608, 1026224]
[h264 @ 0x55d97acc3440] mmco: unref short failure
processed_samples 12901 unjoint_samples 12900 joint_samples 38 [977608, 1026224]
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a13cdd80] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a09bf7c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
processed_samples 13000 unjoint_samples 13000 joint_samples 38 [1047530, 23373]
processed_samples 13000 unjoint_samples 13000 joint_samples 38 [1047530, 23373]
processed_samples 13000 unjoint_samples 13000 joint_samples 39 [1046848, 372449]
processed_samples 13000 unjoint_samples 13000 joint_samples 39 [1046848, 372449]
processed_samples 13000 unjoint_samples 13000 joint_samples 39 [256143, 1040214]
processed_samples 13000 unjoint_samples 13000 joint_samples 39 [256143, 1040214]
processed_samples 13000 unjoint_samples 13000 joint_samples 38 [1046619, 357772]
processed_samples 13000 unjoint_samples 13000 joint_samples 38 [1046619, 357772]
processed_samples 13000 unjoint_samples 13000 joint_samples 38 [1041212, 250793]
processed_samples 13000 unjoint_samples 13000 joint_samples 38 [1041212, 250793]
processed_samples 13000 unjoint_samples 13000 joint_samples 39 [238370, 958193]
processed_samples 13000 unjoint_samples 13000 joint_samples 39 [238370, 958193]
processed_samples 13001 unjoint_samples 13000 joint_samples 39 [1045299, 233945]
processed_samples 13001 unjoint_samples 13000 joint_samples 39 [1045299, 233945]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 13001 unjoint_samples 13000 joint_samples 39 [280622, 1048256]
processed_samples 13001 unjoint_samples 13000 joint_samples 39 [280622, 1048256]
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a09bf7c0] mmco: unref short failure
[h264 @ 0x5592a09bf7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
processed_samples 13100 unjoint_samples 13100 joint_samples 38 [1046619, 642426]
processed_samples 13100 unjoint_samples 13100 joint_samples 39 [530552, 1040214]
processed_samples 13100 unjoint_samples 13100 joint_samples 38 [1047530, 370254]
processed_samples 13100 unjoint_samples 13100 joint_samples 38 [1041212, 523782]
processed_samples 13100 unjoint_samples 13100 joint_samples 39 [527130, 958193]
processed_samples 13100 unjoint_samples 13100 joint_samples 39 [1046848, 828600]
[h264 @ 0x5592a534d800] mmco: unref short failure
processed_samples 13101 unjoint_samples 13100 joint_samples 39 [1045299, 679720]
processed_samples 13100 unjoint_samples 13100 joint_samples 39 [530552, 1040214]
processed_samples 13100 unjoint_samples 13100 joint_samples 38 [1047530, 370254]
processed_samples 13100 unjoint_samples 13100 joint_samples 39 [1046848, 828600]
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a09bf7c0] mmco: unref short failure
[h264 @ 0x5592a09bf7c0] mmco: unref short failure
processed_samples 13101 unjoint_samples 13100 joint_samples 39 [514711, 1048256]
processed_samples 13100 unjoint_samples 13100 joint_samples 38 [1041212, 523782]
processed_samples 13101 unjoint_samples 13100 joint_samples 39 [1045299, 679720]
processed_samples 13100 unjoint_samples 13100 joint_samples 39 [527130, 958193]
processed_samples 13100 unjoint_samples 13100 joint_samples 38 [1046619, 642426]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 13101 unjoint_samples 13100 joint_samples 39 [514711, 1048256]
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a09bf7c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
processed_samples 13200 unjoint_samples 13200 joint_samples 40 [107752, 1040465]
processed_samples 13200 unjoint_samples 13200 joint_samples 39 [857785, 1040214]
processed_samples 13200 unjoint_samples 13200 joint_samples 38 [1047530, 774967]
processed_samples 13200 unjoint_samples 13200 joint_samples 40 [107752, 1040465]
processed_samples 13200 unjoint_samples 13200 joint_samples 38 [1047530, 774967]
processed_samples 13201 unjoint_samples 13200 joint_samples 40 [1047607, 39568]
processed_samples 13200 unjoint_samples 13200 joint_samples 38 [1046619, 996709]
processed_samples 13200 unjoint_samples 13200 joint_samples 39 [900724, 958193]
processed_samples 13200 unjoint_samples 13200 joint_samples 38 [1046619, 996709]
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
processed_samples 13200 unjoint_samples 13200 joint_samples 38 [1041212, 809616]
processed_samples 13200 unjoint_samples 13200 joint_samples 38 [1041212, 809616]
processed_samples 13201 unjoint_samples 13200 joint_samples 40 [1047607, 39568]
processed_samples 13200 unjoint_samples 13200 joint_samples 39 [900724, 958193]
[h264 @ 0x55d97e662e80] mmco: unref short failure
processed_samples 13200 unjoint_samples 13200 joint_samples 39 [857785, 1040214]
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
processed_samples 13201 unjoint_samples 13200 joint_samples 39 [811948, 1048256]
processed_samples 13201 unjoint_samples 13200 joint_samples 39 [811948, 1048256]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c2ff6c0] mmco: unref short failure
[h264 @ 0x55d97c2ff6c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
processed_samples 13300 unjoint_samples 13300 joint_samples 40 [1046102, 134846]
processed_samples 13300 unjoint_samples 13300 joint_samples 40 [1046102, 134846]
processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1047530, 64531]
processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1047530, 64531]
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
processed_samples 13300 unjoint_samples 13300 joint_samples 39 [62185, 1044795]
processed_samples 13300 unjoint_samples 13300 joint_samples 39 [62185, 1044795]
processed_samples 13300 unjoint_samples 13300 joint_samples 40 [446499, 1040465]
processed_samples 13300 unjoint_samples 13300 joint_samples 40 [446499, 1040465]
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1048022, 333463]
processed_samples 13300 unjoint_samples 13300 joint_samples 39 [1048022, 333463]
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1047085, 72331]
processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1047085, 72331]
processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1047607, 418662]
processed_samples 13301 unjoint_samples 13300 joint_samples 40 [1047607, 418662]
processed_samples 13300 unjoint_samples 13300 joint_samples 40 [265210, 1042563]
processed_samples 13300 unjoint_samples 13300 joint_samples 40 [265210, 1042563]
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
processed_samples 13400 unjoint_samples 13400 joint_samples 39 [392912, 1044795]
processed_samples 13400 unjoint_samples 13400 joint_samples 40 [1046102, 466725]
processed_samples 13400 unjoint_samples 13400 joint_samples 40 [724997, 1040465]
processed_samples 13400 unjoint_samples 13400 joint_samples 39 [1047530, 359259]
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 13400 unjoint_samples 13400 joint_samples 39 [1048022, 623524]
processed_samples 13400 unjoint_samples 13400 joint_samples 40 [530262, 1042563]
processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1047085, 378990]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1047607, 698899]
processed_samples 13400 unjoint_samples 13400 joint_samples 39 [392912, 1044795]
processed_samples 13400 unjoint_samples 13400 joint_samples 40 [1046102, 466725]
processed_samples 13400 unjoint_samples 13400 joint_samples 39 [1047530, 359259]
processed_samples 13400 unjoint_samples 13400 joint_samples 40 [724997, 1040465]
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
processed_samples 13400 unjoint_samples 13400 joint_samples 40 [530262, 1042563]
processed_samples 13400 unjoint_samples 13400 joint_samples 39 [1048022, 623524]
processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1047085, 378990]
processed_samples 13401 unjoint_samples 13400 joint_samples 40 [1047607, 698899]
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97d7f2fc0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
processed_samples 13500 unjoint_samples 13500 joint_samples 39 [1047530, 636142]
processed_samples 13500 unjoint_samples 13500 joint_samples 39 [707728, 1044795]
processed_samples 13500 unjoint_samples 13500 joint_samples 39 [1048022, 983195]
processed_samples 13500 unjoint_samples 13500 joint_samples 40 [880860, 1042563]
processed_samples 13500 unjoint_samples 13500 joint_samples 39 [1047530, 636142]
processed_samples 13500 unjoint_samples 13500 joint_samples 40 [1046102, 753256]
processed_samples 13500 unjoint_samples 13500 joint_samples 40 [1011502, 1040465]
processed_samples 13500 unjoint_samples 13500 joint_samples 40 [1011502, 1040465]
processed_samples 13500 unjoint_samples 13500 joint_samples 39 [707728, 1044795]
processed_samples 13500 unjoint_samples 13500 joint_samples 40 [1046102, 753256]
processed_samples 13500 unjoint_samples 13500 joint_samples 40 [880860, 1042563]
processed_samples 13500 unjoint_samples 13500 joint_samples 39 [1048022, 983195]
processed_samples 13501 unjoint_samples 13500 joint_samples 40 [1047085, 700894]
processed_samples 13501 unjoint_samples 13500 joint_samples 40 [1047607, 991283]
processed_samples 13501 unjoint_samples 13500 joint_samples 40 [1047085, 700894]
processed_samples 13501 unjoint_samples 13500 joint_samples 40 [1047607, 991283]
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97c1b3ac0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d980a46440] mmco: unref short failure
[h264 @ 0x55d980a46440] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97b9e9100] Missing reference picture, default is 65530
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] Missing reference picture, default is 65530
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
processed_samples 13600 unjoint_samples 13600 joint_samples 41 [1039553, 114563]
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
processed_samples 13600 unjoint_samples 13600 joint_samples 41 [273046, 1046007]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 13600 unjoint_samples 13600 joint_samples 39 [1047530, 924006]
processed_samples 13600 unjoint_samples 13600 joint_samples 40 [1048022, 328809]
processed_samples 13600 unjoint_samples 13600 joint_samples 40 [1046102, 1040502]
processed_samples 13601 unjoint_samples 13600 joint_samples 41 [240361, 1044179]
processed_samples 13600 unjoint_samples 13600 joint_samples 39 [991109, 1044795]
[h264 @ 0x55d97d7af000] mmco: unref short failure
processed_samples 13601 unjoint_samples 13600 joint_samples 40 [1047085, 911488]
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
processed_samples 13600 unjoint_samples 13600 joint_samples 41 [1039553, 114563]
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 13600 unjoint_samples 13600 joint_samples 41 [273046, 1046007]
processed_samples 13600 unjoint_samples 13600 joint_samples 40 [1048022, 328809]
processed_samples 13600 unjoint_samples 13600 joint_samples 40 [1046102, 1040502]
processed_samples 13601 unjoint_samples 13600 joint_samples 41 [240361, 1044179]
processed_samples 13600 unjoint_samples 13600 joint_samples 39 [991109, 1044795]
processed_samples 13600 unjoint_samples 13600 joint_samples 39 [1047530, 924006]
processed_samples 13601 unjoint_samples 13600 joint_samples 40 [1047085, 911488]
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a2b5cac0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1039553, 382673]
processed_samples 13700 unjoint_samples 13700 joint_samples 40 [1040328, 268071]
processed_samples 13700 unjoint_samples 13700 joint_samples 40 [212262, 1046810]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1046102, 351065]
processed_samples 13700 unjoint_samples 13700 joint_samples 41 [629686, 1046007]
processed_samples 13701 unjoint_samples 13700 joint_samples 41 [367483, 978261]
processed_samples 13700 unjoint_samples 13700 joint_samples 40 [1048022, 659178]
[h264 @ 0x55d97ea8dc80] mmco: unref short failure
[h264 @ 0x55d97ea8dc80] mmco: unref short failure
processed_samples 13701 unjoint_samples 13700 joint_samples 41 [638325, 1044179]
processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1039553, 382673]
processed_samples 13700 unjoint_samples 13700 joint_samples 40 [212262, 1046810]
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
processed_samples 13700 unjoint_samples 13700 joint_samples 40 [1040328, 268071]
processed_samples 13700 unjoint_samples 13700 joint_samples 41 [629686, 1046007]
processed_samples 13700 unjoint_samples 13700 joint_samples 41 [1046102, 351065]
processed_samples 13701 unjoint_samples 13700 joint_samples 41 [367483, 978261]
processed_samples 13700 unjoint_samples 13700 joint_samples 40 [1048022, 659178]
processed_samples 13701 unjoint_samples 13700 joint_samples 41 [638325, 1044179]
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1046102, 623191]
processed_samples 13800 unjoint_samples 13800 joint_samples 40 [621627, 1046810]
processed_samples 13800 unjoint_samples 13800 joint_samples 40 [1040328, 596974]
processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1039553, 724123]
processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1018621, 1046007]
processed_samples 13800 unjoint_samples 13800 joint_samples 40 [1048022, 993655]
processed_samples 13801 unjoint_samples 13800 joint_samples 41 [692426, 978261]
processed_samples 13801 unjoint_samples 13800 joint_samples 41 [968917, 1044179]
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 13800 unjoint_samples 13800 joint_samples 40 [1040328, 596974]
processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1046102, 623191]
processed_samples 13800 unjoint_samples 13800 joint_samples 40 [621627, 1046810]
processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1039553, 724123]
processed_samples 13800 unjoint_samples 13800 joint_samples 41 [1018621, 1046007]
processed_samples 13801 unjoint_samples 13800 joint_samples 41 [692426, 978261]
processed_samples 13801 unjoint_samples 13800 joint_samples 41 [968917, 1044179]
processed_samples 13800 unjoint_samples 13800 joint_samples 40 [1048022, 993655]
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 13900 unjoint_samples 13900 joint_samples 42 [71407, 1044891]
processed_samples 13900 unjoint_samples 13900 joint_samples 42 [233476, 1046982]
processed_samples 13900 unjoint_samples 13900 joint_samples 41 [1046102, 967557]
processed_samples 13900 unjoint_samples 13900 joint_samples 41 [180724, 1046100]
processed_samples 13900 unjoint_samples 13900 joint_samples 40 [1040328, 839344]
processed_samples 13901 unjoint_samples 13900 joint_samples 42 [275966, 1046267]
processed_samples 13900 unjoint_samples 13900 joint_samples 40 [938002, 1046810]
processed_samples 13901 unjoint_samples 13900 joint_samples 41 [961668, 978261]
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
processed_samples 13900 unjoint_samples 13900 joint_samples 42 [71407, 1044891]
processed_samples 13900 unjoint_samples 13900 joint_samples 42 [233476, 1046982]
processed_samples 13900 unjoint_samples 13900 joint_samples 41 [1046102, 967557]
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
processed_samples 13900 unjoint_samples 13900 joint_samples 41 [180724, 1046100]
processed_samples 13901 unjoint_samples 13900 joint_samples 42 [275966, 1046267]
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 13900 unjoint_samples 13900 joint_samples 40 [1040328, 839344]
processed_samples 13900 unjoint_samples 13900 joint_samples 40 [938002, 1046810]
processed_samples 13901 unjoint_samples 13900 joint_samples 41 [961668, 978261]
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
processed_samples 14000 unjoint_samples 14000 joint_samples 42 [199639, 1046780]
processed_samples 14000 unjoint_samples 14000 joint_samples 41 [134019, 1017027]
processed_samples 14000 unjoint_samples 14000 joint_samples 41 [1043157, 247225]
processed_samples 14000 unjoint_samples 14000 joint_samples 42 [528838, 1044891]
processed_samples 14000 unjoint_samples 14000 joint_samples 42 [534717, 1046982]
processed_samples 14001 unjoint_samples 14000 joint_samples 42 [172434, 1037543]
processed_samples 14001 unjoint_samples 14000 joint_samples 42 [736101, 1046267]
processed_samples 14000 unjoint_samples 14000 joint_samples 41 [634079, 1046100]
[h264 @ 0x5592a31fac00] mmco: unref short failure
processed_samples 14000 unjoint_samples 14000 joint_samples 41 [134019, 1017027]
processed_samples 14000 unjoint_samples 14000 joint_samples 42 [199639, 1046780]
processed_samples 14000 unjoint_samples 14000 joint_samples 41 [1043157, 247225]
processed_samples 14000 unjoint_samples 14000 joint_samples 42 [528838, 1044891]
processed_samples 14001 unjoint_samples 14000 joint_samples 42 [172434, 1037543]
processed_samples 14000 unjoint_samples 14000 joint_samples 42 [534717, 1046982]
processed_samples 14001 unjoint_samples 14000 joint_samples 42 [736101, 1046267]
processed_samples 14000 unjoint_samples 14000 joint_samples 41 [634079, 1046100]
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 14100 unjoint_samples 14100 joint_samples 41 [1043157, 665607]
processed_samples 14100 unjoint_samples 14100 joint_samples 42 [524270, 1046780]
processed_samples 14100 unjoint_samples 14100 joint_samples 41 [419570, 1017027]
processed_samples 14101 unjoint_samples 14100 joint_samples 42 [409061, 1037543]
processed_samples 14101 unjoint_samples 14100 joint_samples 43 [13531, 1048354]
processed_samples 14101 unjoint_samples 14100 joint_samples 43 [13531, 1048354]
processed_samples 14100 unjoint_samples 14100 joint_samples 41 [950286, 1046100]
processed_samples 14100 unjoint_samples 14100 joint_samples 41 [419570, 1017027]
processed_samples 14101 unjoint_samples 14100 joint_samples 42 [409061, 1037543]
processed_samples 14100 unjoint_samples 14100 joint_samples 41 [1043157, 665607]
processed_samples 14100 unjoint_samples 14100 joint_samples 42 [524270, 1046780]
processed_samples 14100 unjoint_samples 14100 joint_samples 42 [821853, 1044891]
processed_samples 14100 unjoint_samples 14100 joint_samples 42 [821853, 1044891]
processed_samples 14100 unjoint_samples 14100 joint_samples 42 [797459, 1046982]
[h264 @ 0x55d98337cc80] mmco: unref short failure
processed_samples 14100 unjoint_samples 14100 joint_samples 42 [797459, 1046982]
processed_samples 14100 unjoint_samples 14100 joint_samples 41 [950286, 1046100]
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1cd1c80] mmco: unref short failure
[h264 @ 0x5592a1cd1c80] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a1cd1c80] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a4a36d40] mmco: unref short failure
[h264 @ 0x5592a4a36d40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] [h264 @ 0x5592a1f2c740] mmco: unref short failure
mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
processed_samples 14200 unjoint_samples 14200 joint_samples 43 [1045397, 74108]
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
processed_samples 14200 unjoint_samples 14200 joint_samples 41 [1043157, 1026006]
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
processed_samples 14200 unjoint_samples 14200 joint_samples 43 [1045125, 8073]
processed_samples 14200 unjoint_samples 14200 joint_samples 43 [1045397, 74108]
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
processed_samples 14200 unjoint_samples 14200 joint_samples 41 [1043157, 1026006]
processed_samples 14200 unjoint_samples 14200 joint_samples 43 [1045125, 8073]
processed_samples 14200 unjoint_samples 14200 joint_samples 42 [1032874, 382365]
processed_samples 14200 unjoint_samples 14200 joint_samples 42 [1032874, 382365]
processed_samples 14200 unjoint_samples 14200 joint_samples 42 [1002742, 1046780]
processed_samples 14200 unjoint_samples 14200 joint_samples 42 [1002742, 1046780]
processed_samples 14201 unjoint_samples 14200 joint_samples 43 [363270, 1048354]
processed_samples 14200 unjoint_samples 14200 joint_samples 41 [716281, 1017027]
processed_samples 14200 unjoint_samples 14200 joint_samples 41 [716281, 1017027]
processed_samples 14201 unjoint_samples 14200 joint_samples 43 [363270, 1048354]
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
processed_samples 14201 unjoint_samples 14200 joint_samples 42 [763782, 1037543]
processed_samples 14201 unjoint_samples 14200 joint_samples 42 [763782, 1037543]
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
processed_samples 14300 unjoint_samples 14300 joint_samples 43 [403889, 1046780]
processed_samples 14300 unjoint_samples 14300 joint_samples 43 [403889, 1046780]
processed_samples 14300 unjoint_samples 14300 joint_samples 42 [1045299, 271047]
processed_samples 14300 unjoint_samples 14300 joint_samples 42 [1045299, 271047]
processed_samples 14300 unjoint_samples 14300 joint_samples 43 [1045125, 245824]
processed_samples 14300 unjoint_samples 14300 joint_samples 43 [1045397, 324806]
processed_samples 14300 unjoint_samples 14300 joint_samples 43 [1045125, 245824]
processed_samples 14300 unjoint_samples 14300 joint_samples 43 [1045397, 324806]
processed_samples 14300 unjoint_samples 14300 joint_samples 42 [1032874, 653574]
processed_samples 14300 unjoint_samples 14300 joint_samples 42 [1032874, 653574]
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
processed_samples 14301 unjoint_samples 14300 joint_samples 43 [657009, 1048354]
processed_samples 14301 unjoint_samples 14300 joint_samples 43 [657009, 1048354]
processed_samples 14301 unjoint_samples 14300 joint_samples 43 [1038825, 42053]
processed_samples 14301 unjoint_samples 14300 joint_samples 43 [1038825, 42053]
processed_samples 14300 unjoint_samples 14300 joint_samples 41 [1039424, 1041776]
processed_samples 14300 unjoint_samples 14300 joint_samples 41 [1039424, 1041776]
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97aa0b980] mmco: unref short failure
[h264 @ 0x55d97aa0b980] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
processed_samples 14400 unjoint_samples 14400 joint_samples 42 [1045299, 559025]
processed_samples 14400 unjoint_samples 14400 joint_samples 42 [426135, 1041776]
[h264 @ 0x5592a1f04380] mmco: unref short failure
processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1045125, 488955]
processed_samples 14400 unjoint_samples 14400 joint_samples 43 [775376, 1046780]
processed_samples 14401 unjoint_samples 14400 joint_samples 43 [1038825, 313572]
processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1045397, 701148]
processed_samples 14400 unjoint_samples 14400 joint_samples 42 [1046738, 1040413]
processed_samples 14401 unjoint_samples 14400 joint_samples 43 [1005725, 1048354]
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
processed_samples 14400 unjoint_samples 14400 joint_samples 42 [426135, 1041776]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 14400 unjoint_samples 14400 joint_samples 42 [1045299, 559025]
processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1045125, 488955]
processed_samples 14400 unjoint_samples 14400 joint_samples 43 [775376, 1046780]
processed_samples 14400 unjoint_samples 14400 joint_samples 43 [1045397, 701148]
processed_samples 14401 unjoint_samples 14400 joint_samples 43 [1038825, 313572]
processed_samples 14401 unjoint_samples 14400 joint_samples 43 [1005725, 1048354]
processed_samples 14400 unjoint_samples 14400 joint_samples 42 [1046738, 1040413]
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
processed_samples 14500 unjoint_samples 14500 joint_samples 44 [123723, 977942]
processed_samples 14500 unjoint_samples 14500 joint_samples 44 [1031004, 93887]
processed_samples 14500 unjoint_samples 14500 joint_samples 42 [704008, 1041776]
processed_samples 14500 unjoint_samples 14500 joint_samples 42 [1045299, 954908]
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
processed_samples 14500 unjoint_samples 14500 joint_samples 43 [280438, 1040413]
processed_samples 14501 unjoint_samples 14500 joint_samples 43 [1038825, 613701]
processed_samples 14500 unjoint_samples 14500 joint_samples 44 [123723, 977942]
processed_samples 14501 unjoint_samples 14500 joint_samples 44 [1032323, 301091]
processed_samples 14500 unjoint_samples 14500 joint_samples 43 [1045125, 742310]
processed_samples 14500 unjoint_samples 14500 joint_samples 44 [1031004, 93887]
processed_samples 14500 unjoint_samples 14500 joint_samples 42 [1045299, 954908]
processed_samples 14500 unjoint_samples 14500 joint_samples 43 [280438, 1040413]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 14501 unjoint_samples 14500 joint_samples 44 [1032323, 301091]
processed_samples 14500 unjoint_samples 14500 joint_samples 42 [704008, 1041776]
processed_samples 14501 unjoint_samples 14500 joint_samples 43 [1038825, 613701]
processed_samples 14500 unjoint_samples 14500 joint_samples 43 [1045125, 742310]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
processed_samples 14600 unjoint_samples 14600 joint_samples 44 [592104, 977942]
processed_samples 14600 unjoint_samples 14600 joint_samples 43 [1045299, 324354]
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
processed_samples 14600 unjoint_samples 14600 joint_samples 44 [345015, 976518]
processed_samples 14600 unjoint_samples 14600 joint_samples 44 [1031004, 392697]
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
processed_samples 14600 unjoint_samples 14600 joint_samples 42 [951652, 1041776]
processed_samples 14600 unjoint_samples 14600 joint_samples 43 [537651, 1040413]
processed_samples 14600 unjoint_samples 14600 joint_samples 44 [592104, 977942]
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 14600 unjoint_samples 14600 joint_samples 43 [1045299, 324354]
processed_samples 14601 unjoint_samples 14600 joint_samples 44 [1032323, 520737]
processed_samples 14600 unjoint_samples 14600 joint_samples 44 [345015, 976518]
processed_samples 14600 unjoint_samples 14600 joint_samples 44 [1031004, 392697]
processed_samples 14601 unjoint_samples 14600 joint_samples 43 [1038825, 909005]
processed_samples 14600 unjoint_samples 14600 joint_samples 42 [951652, 1041776]
processed_samples 14600 unjoint_samples 14600 joint_samples 43 [537651, 1040413]
processed_samples 14601 unjoint_samples 14600 joint_samples 44 [1032323, 520737]
processed_samples 14601 unjoint_samples 14600 joint_samples 43 [1038825, 909005]
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 14700 unjoint_samples 14700 joint_samples 44 [852174, 977942]
processed_samples 14700 unjoint_samples 14700 joint_samples 43 [1044934, 229681]
processed_samples 14700 unjoint_samples 14700 joint_samples 44 [1031004, 744884]
processed_samples 14700 unjoint_samples 14700 joint_samples 43 [1045299, 546979]
processed_samples 14700 unjoint_samples 14700 joint_samples 44 [696847, 976518]
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 14701 unjoint_samples 14700 joint_samples 44 [1032323, 1020422]
processed_samples 14701 unjoint_samples 14700 joint_samples 44 [1046351, 147027]
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
processed_samples 14700 unjoint_samples 14700 joint_samples 43 [851979, 1040413]
processed_samples 14700 unjoint_samples 14700 joint_samples 44 [852174, 977942]
processed_samples 14700 unjoint_samples 14700 joint_samples 43 [1044934, 229681]
processed_samples 14701 unjoint_samples 14700 joint_samples 44 [1046351, 147027]
processed_samples 14700 unjoint_samples 14700 joint_samples 43 [1045299, 546979]
processed_samples 14700 unjoint_samples 14700 joint_samples 44 [696847, 976518]
processed_samples 14700 unjoint_samples 14700 joint_samples 44 [1031004, 744884]
processed_samples 14701 unjoint_samples 14700 joint_samples 44 [1032323, 1020422]
processed_samples 14700 unjoint_samples 14700 joint_samples 43 [851979, 1040413]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4bc4600] mmco: unref short failure
[h264 @ 0x5592a4bc4600] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 14800 unjoint_samples 14800 joint_samples 45 [1041287, 29434]
processed_samples 14800 unjoint_samples 14800 joint_samples 45 [1041287, 29434]
processed_samples 14800 unjoint_samples 14800 joint_samples 44 [1046074, 69014]
processed_samples 14800 unjoint_samples 14800 joint_samples 44 [1046074, 69014]
processed_samples 14800 unjoint_samples 14800 joint_samples 43 [1044934, 556350]
processed_samples 14800 unjoint_samples 14800 joint_samples 43 [1044934, 556350]
processed_samples 14800 unjoint_samples 14800 joint_samples 43 [1045299, 777062]
processed_samples 14801 unjoint_samples 14800 joint_samples 45 [1032462, 253539]
processed_samples 14801 unjoint_samples 14800 joint_samples 45 [1032462, 253539]
processed_samples 14800 unjoint_samples 14800 joint_samples 43 [1045299, 777062]
processed_samples 14801 unjoint_samples 14800 joint_samples 44 [1046351, 563591]
processed_samples 14800 unjoint_samples 14800 joint_samples 44 [1033303, 1031746]
processed_samples 14801 unjoint_samples 14800 joint_samples 44 [1046351, 563591]
processed_samples 14800 unjoint_samples 14800 joint_samples 44 [1033303, 1031746]
processed_samples 14800 unjoint_samples 14800 joint_samples 44 [984343, 979265]
processed_samples 14800 unjoint_samples 14800 joint_samples 44 [984343, 979265]
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a4bc4600] mmco: unref short failure
[h264 @ 0x5592a4bc4600] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 14900 unjoint_samples 14900 joint_samples 44 [73299, 1044564]
processed_samples 14900 unjoint_samples 14900 joint_samples 44 [73299, 1044564]
processed_samples 14900 unjoint_samples 14900 joint_samples 45 [277812, 1028174]
processed_samples 14900 unjoint_samples 14900 joint_samples 45 [277812, 1028174]
processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1045145, 328300]
processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1041287, 499091]
processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1045145, 328300]
processed_samples 14900 unjoint_samples 14900 joint_samples 45 [1041287, 499091]
processed_samples 14901 unjoint_samples 14900 joint_samples 45 [1046351, 18612]
processed_samples 14901 unjoint_samples 14900 joint_samples 45 [1046351, 18612]
processed_samples 14901 unjoint_samples 14900 joint_samples 45 [1032462, 545615]
processed_samples 14900 unjoint_samples 14900 joint_samples 44 [1046074, 435653]
processed_samples 14900 unjoint_samples 14900 joint_samples 44 [1046074, 435653]
processed_samples 14901 unjoint_samples 14900 joint_samples 45 [1032462, 545615]
processed_samples 14900 unjoint_samples 14900 joint_samples 43 [1044934, 907756]
processed_samples 14900 unjoint_samples 14900 joint_samples 43 [1044934, 907756]
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ee15800] mmco: unref short failure
[h264 @ 0x55d97ee15800] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e9c79c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e9c79c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
processed_samples 15000 unjoint_samples 15000 joint_samples 44 [423982, 1044564]
processed_samples 15000 unjoint_samples 15000 joint_samples 44 [1047194, 194371]
processed_samples 15000 unjoint_samples 15000 joint_samples 44 [1046074, 730185]
processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1045145, 636085]
processed_samples 15000 unjoint_samples 15000 joint_samples 45 [521765, 1028174]
processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1041287, 943366]
processed_samples 15001 unjoint_samples 15000 joint_samples 45 [1046351, 348296]
processed_samples 15001 unjoint_samples 15000 joint_samples 45 [1032462, 834614]
[h264 @ 0x5592a2a54580] mmco: unref short failure
processed_samples 15000 unjoint_samples 15000 joint_samples 44 [423982, 1044564]
processed_samples 15000 unjoint_samples 15000 joint_samples 44 [1047194, 194371]
[h264 @ 0x5592a105f780] mmco: unref short failure
processed_samples 15000 unjoint_samples 15000 joint_samples 44 [1046074, 730185]
processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1045145, 636085]
processed_samples 15000 unjoint_samples 15000 joint_samples 45 [521765, 1028174]
processed_samples 15001 unjoint_samples 15000 joint_samples 45 [1046351, 348296]
processed_samples 15000 unjoint_samples 15000 joint_samples 45 [1041287, 943366]
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
processed_samples 15001 unjoint_samples 15000 joint_samples 45 [1032462, 834614]
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a2362480] mmco: unref short failure
[h264 @ 0x5592a2362480] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
processed_samples 15100 unjoint_samples 15100 joint_samples 46 [314433, 1045684]
[h264 @ 0x55d97b3130c0] mmco: unref short failure
processed_samples 15100 unjoint_samples 15100 joint_samples 45 [1045145, 1015127]
processed_samples 15100 unjoint_samples 15100 joint_samples 44 [704160, 1044564]
processed_samples 15100 unjoint_samples 15100 joint_samples 44 [1046074, 968626]
processed_samples 15100 unjoint_samples 15100 joint_samples 44 [1047194, 659435]
processed_samples 15100 unjoint_samples 15100 joint_samples 45 [826786, 1028174]
processed_samples 15101 unjoint_samples 15100 joint_samples 46 [232866, 1047558]
processed_samples 15101 unjoint_samples 15100 joint_samples 45 [1046351, 663134]
processed_samples 15100 unjoint_samples 15100 joint_samples 45 [826786, 1028174]
[h264 @ 0x5592a48aaa80] mmco: unref short failure
processed_samples 15100 unjoint_samples 15100 joint_samples 46 [314433, 1045684]
processed_samples 15100 unjoint_samples 15100 joint_samples 44 [704160, 1044564]
processed_samples 15100 unjoint_samples 15100 joint_samples 44 [1047194, 659435]
processed_samples 15100 unjoint_samples 15100 joint_samples 44 [1046074, 968626]
processed_samples 15100 unjoint_samples 15100 joint_samples 45 [1045145, 1015127]
processed_samples 15101 unjoint_samples 15100 joint_samples 46 [232866, 1047558]
processed_samples 15101 unjoint_samples 15100 joint_samples 45 [1046351, 663134]
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97b9674c0] mmco: unref short failure
[h264 @ 0x55d97b9674c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x55d97a51b240] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
processed_samples 15200 unjoint_samples 15200 joint_samples 46 [1047285, 204564]
[h264 @ 0x55d982fb5580] mmco: unref short failure
processed_samples 15200 unjoint_samples 15200 joint_samples 46 [1044821, 86844]
processed_samples 15200 unjoint_samples 15200 joint_samples 46 [717585, 1045684]
processed_samples 15200 unjoint_samples 15200 joint_samples 45 [1046113, 186576]
processed_samples 15200 unjoint_samples 15200 joint_samples 44 [1047194, 1011148]
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
processed_samples 15200 unjoint_samples 15200 joint_samples 46 [1047285, 204564]
[h264 @ 0x5592a118ff40] mmco: unref short failure
processed_samples 15200 unjoint_samples 15200 joint_samples 45 [1046113, 186576]
processed_samples 15200 unjoint_samples 15200 joint_samples 46 [717585, 1045684]
processed_samples 15200 unjoint_samples 15200 joint_samples 46 [1044821, 86844]
processed_samples 15200 unjoint_samples 15200 joint_samples 44 [923221, 1044564]
processed_samples 15201 unjoint_samples 15200 joint_samples 46 [488456, 1047558]
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
processed_samples 15200 unjoint_samples 15200 joint_samples 44 [1047194, 1011148]
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
processed_samples 15201 unjoint_samples 15200 joint_samples 46 [732557, 731544]
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
processed_samples 15201 unjoint_samples 15200 joint_samples 46 [488456, 1047558]
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
processed_samples 15200 unjoint_samples 15200 joint_samples 44 [923221, 1044564]
processed_samples 15201 unjoint_samples 15200 joint_samples 46 [732557, 731544]
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a2c76200] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97e9cdf80] mmco: unref short failure
[h264 @ 0x55d97e9cdf80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e9cdf80] mmco: unref short failure
[h264 @ 0x55d97e9cdf80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e9cdf80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
processed_samples 15300 unjoint_samples 15300 joint_samples 47 [1033134, 49104]
processed_samples 15300 unjoint_samples 15300 joint_samples 45 [1047585, 146793]
processed_samples 15300 unjoint_samples 15300 joint_samples 45 [291073, 1046659]
processed_samples 15300 unjoint_samples 15300 joint_samples 47 [1033134, 49104]
processed_samples 15300 unjoint_samples 15300 joint_samples 46 [1047285, 523314]
processed_samples 15300 unjoint_samples 15300 joint_samples 45 [1047585, 146793]
processed_samples 15300 unjoint_samples 15300 joint_samples 45 [291073, 1046659]
processed_samples 15300 unjoint_samples 15300 joint_samples 46 [1044821, 416307]
processed_samples 15300 unjoint_samples 15300 joint_samples 46 [1047285, 523314]
processed_samples 15300 unjoint_samples 15300 joint_samples 46 [1044821, 416307]
processed_samples 15300 unjoint_samples 15300 joint_samples 45 [1046113, 570609]
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
processed_samples 15300 unjoint_samples 15300 joint_samples 45 [1046113, 570609]
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
processed_samples 15301 unjoint_samples 15300 joint_samples 46 [914473, 937137]
processed_samples 15301 unjoint_samples 15300 joint_samples 46 [774691, 1047558]
processed_samples 15301 unjoint_samples 15300 joint_samples 46 [914473, 937137]
processed_samples 15301 unjoint_samples 15300 joint_samples 46 [774691, 1047558]
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
processed_samples 15400 unjoint_samples 15400 joint_samples 45 [1047585, 548083]
processed_samples 15400 unjoint_samples 15400 joint_samples 47 [1033134, 308411]
processed_samples 15400 unjoint_samples 15400 joint_samples 45 [644295, 1046659]
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
processed_samples 15400 unjoint_samples 15400 joint_samples 46 [1044821, 679050]
processed_samples 15401 unjoint_samples 15400 joint_samples 47 [1022290, 127160]
processed_samples 15401 unjoint_samples 15400 joint_samples 47 [123748, 1047558]
processed_samples 15400 unjoint_samples 15400 joint_samples 46 [1047285, 971328]
processed_samples 15400 unjoint_samples 15400 joint_samples 45 [1046113, 838243]
[h264 @ 0x55d97e32fd00] mmco: unref short failure
processed_samples 15400 unjoint_samples 15400 joint_samples 46 [1044821, 679050]
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
processed_samples 15400 unjoint_samples 15400 joint_samples 45 [1047585, 548083]
processed_samples 15400 unjoint_samples 15400 joint_samples 47 [1033134, 308411]
processed_samples 15400 unjoint_samples 15400 joint_samples 45 [644295, 1046659]
processed_samples 15400 unjoint_samples 15400 joint_samples 45 [1046113, 838243]
processed_samples 15401 unjoint_samples 15400 joint_samples 47 [1022290, 127160]
processed_samples 15401 unjoint_samples 15400 joint_samples 47 [123748, 1047558]
processed_samples 15400 unjoint_samples 15400 joint_samples 46 [1047285, 971328]
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
processed_samples 15500 unjoint_samples 15500 joint_samples 46 [1047040, 54955]
processed_samples 15500 unjoint_samples 15500 joint_samples 47 [91263, 1045086]
processed_samples 15500 unjoint_samples 15500 joint_samples 47 [249631, 1046373]
processed_samples 15500 unjoint_samples 15500 joint_samples 46 [1047040, 54955]
processed_samples 15500 unjoint_samples 15500 joint_samples 47 [91263, 1045086]
processed_samples 15500 unjoint_samples 15500 joint_samples 47 [249631, 1046373]
processed_samples 15500 unjoint_samples 15500 joint_samples 47 [1033134, 546513]
processed_samples 15500 unjoint_samples 15500 joint_samples 47 [1033134, 546513]
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
processed_samples 15501 unjoint_samples 15500 joint_samples 47 [1022290, 395840]
processed_samples 15501 unjoint_samples 15500 joint_samples 47 [416558, 1047558]
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
processed_samples 15500 unjoint_samples 15500 joint_samples 45 [946163, 1046659]
processed_samples 15501 unjoint_samples 15500 joint_samples 47 [1022290, 395840]
processed_samples 15500 unjoint_samples 15500 joint_samples 45 [1047585, 894363]
processed_samples 15500 unjoint_samples 15500 joint_samples 45 [946163, 1046659]
processed_samples 15501 unjoint_samples 15500 joint_samples 47 [416558, 1047558]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
processed_samples 15500 unjoint_samples 15500 joint_samples 45 [1047585, 894363]
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
processed_samples 15600 unjoint_samples 15600 joint_samples 47 [433552, 1045086]
processed_samples 15600 unjoint_samples 15600 joint_samples 46 [1047585, 89725]
processed_samples 15600 unjoint_samples 15600 joint_samples 46 [1020751, 262251]
processed_samples 15600 unjoint_samples 15600 joint_samples 46 [1047040, 420769]
processed_samples 15600 unjoint_samples 15600 joint_samples 47 [1033134, 832648]
processed_samples 15600 unjoint_samples 15600 joint_samples 47 [550452, 1046373]
processed_samples 15601 unjoint_samples 15600 joint_samples 47 [1022290, 752584]
processed_samples 15601 unjoint_samples 15600 joint_samples 47 [715444, 1047558]
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
processed_samples 15600 unjoint_samples 15600 joint_samples 46 [1047040, 420769]
processed_samples 15600 unjoint_samples 15600 joint_samples 47 [433552, 1045086]
processed_samples 15600 unjoint_samples 15600 joint_samples 46 [1047585, 89725]
processed_samples 15600 unjoint_samples 15600 joint_samples 46 [1020751, 262251]
processed_samples 15600 unjoint_samples 15600 joint_samples 47 [550452, 1046373]
processed_samples 15601 unjoint_samples 15600 joint_samples 47 [715444, 1047558]
processed_samples 15601 unjoint_samples 15600 joint_samples 47 [1022290, 752584]
processed_samples 15600 unjoint_samples 15600 joint_samples 47 [1033134, 832648]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4a64840] mmco: unref short failure
[h264 @ 0x5592a4a64840] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
processed_samples 15700 unjoint_samples 15700 joint_samples 48 [86226, 1046407]
processed_samples 15700 unjoint_samples 15700 joint_samples 48 [86226, 1046407]
processed_samples 15701 unjoint_samples 15700 joint_samples 48 [1044833, 21531]
processed_samples 15701 unjoint_samples 15700 joint_samples 48 [1044833, 21531]
processed_samples 15700 unjoint_samples 15700 joint_samples 48 [1041476, 49946]
processed_samples 15700 unjoint_samples 15700 joint_samples 48 [1041476, 49946]
processed_samples 15700 unjoint_samples 15700 joint_samples 46 [1047585, 606123]
processed_samples 15700 unjoint_samples 15700 joint_samples 46 [1047585, 606123]
processed_samples 15700 unjoint_samples 15700 joint_samples 46 [1020751, 695790]
processed_samples 15700 unjoint_samples 15700 joint_samples 46 [1020751, 695790]
processed_samples 15700 unjoint_samples 15700 joint_samples 47 [742481, 1045086]
processed_samples 15700 unjoint_samples 15700 joint_samples 47 [742481, 1045086]
processed_samples 15700 unjoint_samples 15700 joint_samples 46 [1047040, 945993]
processed_samples 15701 unjoint_samples 15700 joint_samples 47 [1043324, 1041525]
processed_samples 15700 unjoint_samples 15700 joint_samples 46 [1047040, 945993]
processed_samples 15701 unjoint_samples 15700 joint_samples 47 [1043324, 1041525]
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97aed1e00] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97c0d6c80] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
processed_samples 15800 unjoint_samples 15800 joint_samples 46 [1047585, 916939]
[h264 @ 0x55d97e08af80] mmco: unref short failure
processed_samples 15800 unjoint_samples 15800 joint_samples 48 [827908, 1046407]
processed_samples 15800 unjoint_samples 15800 joint_samples 48 [1041476, 279973]
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
processed_samples 15800 unjoint_samples 15800 joint_samples 47 [1047040, 205206]
processed_samples 15800 unjoint_samples 15800 joint_samples 46 [1020751, 986706]
processed_samples 15801 unjoint_samples 15800 joint_samples 48 [396325, 1041525]
processed_samples 15801 unjoint_samples 15800 joint_samples 48 [1044833, 502152]
processed_samples 15800 unjoint_samples 15800 joint_samples 47 [1047788, 1047361]
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
processed_samples 15800 unjoint_samples 15800 joint_samples 48 [1041476, 279973]
processed_samples 15800 unjoint_samples 15800 joint_samples 46 [1047585, 916939]
[h264 @ 0x5592a0e40300] mmco: unref short failure
processed_samples 15800 unjoint_samples 15800 joint_samples 46 [1020751, 986706]
processed_samples 15801 unjoint_samples 15800 joint_samples 48 [396325, 1041525]
processed_samples 15800 unjoint_samples 15800 joint_samples 47 [1047040, 205206]
processed_samples 15800 unjoint_samples 15800 joint_samples 47 [1047788, 1047361]
processed_samples 15801 unjoint_samples 15800 joint_samples 48 [1044833, 502152]
processed_samples 15800 unjoint_samples 15800 joint_samples 48 [827908, 1046407]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a109e4c0] [h264 @ 0x55d96f79c540] mmco: unref short failure
mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
processed_samples 15900 unjoint_samples 15900 joint_samples 49 [67085, 1047218]
processed_samples 15900 unjoint_samples 15900 joint_samples 47 [278563, 1017735]
[h264 @ 0x55d97f274580] mmco: unref short failure
processed_samples 15900 unjoint_samples 15900 joint_samples 48 [251760, 1047361]
processed_samples 15900 unjoint_samples 15900 joint_samples 47 [1047040, 552071]
processed_samples 15900 unjoint_samples 15900 joint_samples 48 [1041476, 670515]
processed_samples 15900 unjoint_samples 15900 joint_samples 47 [168390, 1028439]
processed_samples 15901 unjoint_samples 15900 joint_samples 48 [757020, 1041525]
processed_samples 15901 unjoint_samples 15900 joint_samples 48 [1044833, 809313]
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
processed_samples 15900 unjoint_samples 15900 joint_samples 49 [67085, 1047218]
processed_samples 15900 unjoint_samples 15900 joint_samples 48 [251760, 1047361]
[h264 @ 0x5592a4deb500] mmco: unref short failure
processed_samples 15900 unjoint_samples 15900 joint_samples 47 [278563, 1017735]
processed_samples 15900 unjoint_samples 15900 joint_samples 47 [168390, 1028439]
processed_samples 15900 unjoint_samples 15900 joint_samples 47 [1047040, 552071]
processed_samples 15900 unjoint_samples 15900 joint_samples 48 [1041476, 670515]
processed_samples 15901 unjoint_samples 15900 joint_samples 48 [757020, 1041525]
processed_samples 15901 unjoint_samples 15900 joint_samples 48 [1044833, 809313]
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a15c3a80] mmco: unref short failure
[h264 @ 0x5592a15c3a80] mmco: unref short failure
[h264 @ 0x559295e768c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
processed_samples 16000 unjoint_samples 16000 joint_samples 47 [1047040, 901688]
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 16001 unjoint_samples 16000 joint_samples 49 [1045045, 44727]
processed_samples 16001 unjoint_samples 16000 joint_samples 49 [34762, 1046595]
processed_samples 16001 unjoint_samples 16000 joint_samples 49 [1045045, 44727]
processed_samples 16001 unjoint_samples 16000 joint_samples 49 [34762, 1046595]
processed_samples 16000 unjoint_samples 16000 joint_samples 48 [628730, 1047361]
processed_samples 16000 unjoint_samples 16000 joint_samples 47 [615945, 1017735]
processed_samples 16000 unjoint_samples 16000 joint_samples 47 [615945, 1017735]
processed_samples 16000 unjoint_samples 16000 joint_samples 47 [493861, 1028439]
processed_samples 16000 unjoint_samples 16000 joint_samples 48 [628730, 1047361]
processed_samples 16000 unjoint_samples 16000 joint_samples 47 [493861, 1028439]
processed_samples 16000 unjoint_samples 16000 joint_samples 47 [1047040, 901688]
processed_samples 16000 unjoint_samples 16000 joint_samples 49 [361426, 1047218]
processed_samples 16000 unjoint_samples 16000 joint_samples 49 [361426, 1047218]
processed_samples 16000 unjoint_samples 16000 joint_samples 48 [1041476, 958259]
processed_samples 16000 unjoint_samples 16000 joint_samples 48 [1041476, 958259]
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 16100 unjoint_samples 16100 joint_samples 47 [819925, 1028439]
processed_samples 16100 unjoint_samples 16100 joint_samples 49 [172832, 1045731]
[h264 @ 0x5592a0d96180] mmco: unref short failure
processed_samples 16100 unjoint_samples 16100 joint_samples 48 [254762, 1046158]
processed_samples 16100 unjoint_samples 16100 joint_samples 49 [721641, 1047218]
processed_samples 16100 unjoint_samples 16100 joint_samples 47 [898774, 1017735]
processed_samples 16100 unjoint_samples 16100 joint_samples 48 [917760, 1047361]
processed_samples 16101 unjoint_samples 16100 joint_samples 49 [1045045, 304232]
processed_samples 16101 unjoint_samples 16100 joint_samples 49 [358733, 1046595]
processed_samples 16100 unjoint_samples 16100 joint_samples 47 [819925, 1028439]
processed_samples 16100 unjoint_samples 16100 joint_samples 49 [172832, 1045731]
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
processed_samples 16100 unjoint_samples 16100 joint_samples 48 [254762, 1046158]
processed_samples 16100 unjoint_samples 16100 joint_samples 47 [898774, 1017735]
processed_samples 16100 unjoint_samples 16100 joint_samples 49 [721641, 1047218]
processed_samples 16100 unjoint_samples 16100 joint_samples 48 [917760, 1047361]
processed_samples 16101 unjoint_samples 16100 joint_samples 49 [358733, 1046595]
processed_samples 16101 unjoint_samples 16100 joint_samples 49 [1045045, 304232]
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a16ec440] mmco: unref short failure
[h264 @ 0x5592a16ec440] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a42f34c0] mmco: unref short failure
[h264 @ 0x5592a42f34c0] mmco: unref short failure
[h264 @ 0x5592a42f34c0] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a16ec440] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
processed_samples 16200 unjoint_samples 16200 joint_samples 49 [466221, 1045731]
processed_samples 16200 unjoint_samples 16200 joint_samples 48 [1035192, 137634]
processed_samples 16201 unjoint_samples 16200 joint_samples 49 [1045045, 622138]
processed_samples 16200 unjoint_samples 16200 joint_samples 48 [289243, 1043817]
processed_samples 16200 unjoint_samples 16200 joint_samples 49 [1046975, 235309]
processed_samples 16201 unjoint_samples 16200 joint_samples 49 [694665, 1046595]
processed_samples 16200 unjoint_samples 16200 joint_samples 48 [545116, 1046158]
processed_samples 16200 unjoint_samples 16200 joint_samples 49 [945013, 1047218]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
[h264 @ 0x55d97e141680] mmco: unref short failure
processed_samples 16200 unjoint_samples 16200 joint_samples 49 [466221, 1045731]
processed_samples 16200 unjoint_samples 16200 joint_samples 48 [1035192, 137634]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 16200 unjoint_samples 16200 joint_samples 49 [1046975, 235309]
processed_samples 16200 unjoint_samples 16200 joint_samples 48 [289243, 1043817]
processed_samples 16200 unjoint_samples 16200 joint_samples 48 [545116, 1046158]
processed_samples 16201 unjoint_samples 16200 joint_samples 49 [1045045, 622138]
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
processed_samples 16201 unjoint_samples 16200 joint_samples 49 [694665, 1046595]
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
processed_samples 16200 unjoint_samples 16200 joint_samples 49 [945013, 1047218]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a16ec440] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a2b5cac0] mmco: unref short failure
[h264 @ 0x5592a2b5cac0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
processed_samples 16300 unjoint_samples 16300 joint_samples 48 [1035192, 427655]
processed_samples 16300 unjoint_samples 16300 joint_samples 48 [1035192, 427655]
processed_samples 16300 unjoint_samples 16300 joint_samples 50 [1044443, 109663]
processed_samples 16300 unjoint_samples 16300 joint_samples 50 [1044443, 109663]
processed_samples 16301 unjoint_samples 16300 joint_samples 50 [12362, 1046595]
processed_samples 16301 unjoint_samples 16300 joint_samples 50 [12362, 1046595]
processed_samples 16300 unjoint_samples 16300 joint_samples 49 [1046975, 538800]
processed_samples 16300 unjoint_samples 16300 joint_samples 49 [1046975, 538800]
processed_samples 16300 unjoint_samples 16300 joint_samples 48 [644206, 1043817]
processed_samples 16300 unjoint_samples 16300 joint_samples 48 [644206, 1043817]
processed_samples 16300 unjoint_samples 16300 joint_samples 49 [682432, 1045731]
processed_samples 16300 unjoint_samples 16300 joint_samples 49 [682432, 1045731]
processed_samples 16301 unjoint_samples 16300 joint_samples 49 [1045045, 890965]
processed_samples 16300 unjoint_samples 16300 joint_samples 48 [860838, 1046158]
processed_samples 16300 unjoint_samples 16300 joint_samples 48 [860838, 1046158]
processed_samples 16301 unjoint_samples 16300 joint_samples 49 [1045045, 890965]
[h264 @ 0x5592a16ec440] mmco: unref short failure
[h264 @ 0x5592a16ec440] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a09d6400] mmco: unref short failure
[h264 @ 0x5592a09d6400] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
processed_samples 16400 unjoint_samples 16400 joint_samples 49 [1043114, 1045731]
processed_samples 16400 unjoint_samples 16400 joint_samples 49 [1043114, 1045731]
processed_samples 16400 unjoint_samples 16400 joint_samples 48 [1035192, 797709]
processed_samples 16400 unjoint_samples 16400 joint_samples 49 [1039244, 104184]
processed_samples 16400 unjoint_samples 16400 joint_samples 49 [1046975, 842559]
processed_samples 16400 unjoint_samples 16400 joint_samples 48 [1035192, 797709]
processed_samples 16400 unjoint_samples 16400 joint_samples 50 [1044443, 483096]
processed_samples 16401 unjoint_samples 16400 joint_samples 50 [1045045, 112814]
processed_samples 16401 unjoint_samples 16400 joint_samples 50 [305771, 1046595]
processed_samples 16400 unjoint_samples 16400 joint_samples 50 [1044443, 483096]
processed_samples 16400 unjoint_samples 16400 joint_samples 49 [1039244, 104184]
processed_samples 16400 unjoint_samples 16400 joint_samples 48 [978490, 1043817]
processed_samples 16400 unjoint_samples 16400 joint_samples 49 [1046975, 842559]
processed_samples 16401 unjoint_samples 16400 joint_samples 50 [1045045, 112814]
processed_samples 16401 unjoint_samples 16400 joint_samples 50 [305771, 1046595]
processed_samples 16400 unjoint_samples 16400 joint_samples 48 [978490, 1043817]
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
processed_samples 16500 unjoint_samples 16500 joint_samples 49 [1046325, 45150]
processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1046627, 378970]
processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1046975, 88099]
processed_samples 16501 unjoint_samples 16500 joint_samples 50 [1045045, 420102]
processed_samples 16501 unjoint_samples 16500 joint_samples 50 [589259, 1046595]
processed_samples 16500 unjoint_samples 16500 joint_samples 49 [1039244, 430760]
processed_samples 16500 unjoint_samples 16500 joint_samples 49 [145439, 1047364]
processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1046627, 378970]
processed_samples 16500 unjoint_samples 16500 joint_samples 49 [1046325, 45150]
processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1044443, 867920]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1046975, 88099]
processed_samples 16500 unjoint_samples 16500 joint_samples 49 [145439, 1047364]
processed_samples 16500 unjoint_samples 16500 joint_samples 50 [1044443, 867920]
processed_samples 16501 unjoint_samples 16500 joint_samples 50 [1045045, 420102]
processed_samples 16500 unjoint_samples 16500 joint_samples 49 [1039244, 430760]
processed_samples 16501 unjoint_samples 16500 joint_samples 50 [589259, 1046595]
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a23c6a40] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97a6a1dc0] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
processed_samples 16600 unjoint_samples 16600 joint_samples 51 [139319, 1022027]
processed_samples 16600 unjoint_samples 16600 joint_samples 49 [1046325, 334650]
processed_samples 16600 unjoint_samples 16600 joint_samples 50 [1046627, 659394]
processed_samples 16600 unjoint_samples 16600 joint_samples 50 [1046975, 510266]
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 16600 unjoint_samples 16600 joint_samples 49 [436467, 1047364]
processed_samples 16601 unjoint_samples 16600 joint_samples 50 [833111, 1046595]
processed_samples 16600 unjoint_samples 16600 joint_samples 51 [139319, 1022027]
processed_samples 16600 unjoint_samples 16600 joint_samples 49 [1046325, 334650]
processed_samples 16600 unjoint_samples 16600 joint_samples 50 [1046975, 510266]
[h264 @ 0x5592a4810c40] mmco: unref short failure
processed_samples 16600 unjoint_samples 16600 joint_samples 49 [436467, 1047364]
processed_samples 16600 unjoint_samples 16600 joint_samples 50 [1046627, 659394]
processed_samples 16600 unjoint_samples 16600 joint_samples 49 [1039244, 706637]
processed_samples 16601 unjoint_samples 16600 joint_samples 50 [833111, 1046595]
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
processed_samples 16601 unjoint_samples 16600 joint_samples 50 [1045045, 775199]
processed_samples 16600 unjoint_samples 16600 joint_samples 49 [1039244, 706637]
[h264 @ 0x5592a0f1f600] mmco: unref short failure
processed_samples 16601 unjoint_samples 16600 joint_samples 50 [1045045, 775199]
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
processed_samples 16701 unjoint_samples 16700 joint_samples 51 [1045045, 29138]
processed_samples 16701 unjoint_samples 16700 joint_samples 51 [58471, 1046595]
processed_samples 16701 unjoint_samples 16700 joint_samples 51 [1045045, 29138]
processed_samples 16701 unjoint_samples 16700 joint_samples 51 [58471, 1046595]
processed_samples 16700 unjoint_samples 16700 joint_samples 51 [390215, 1022027]
processed_samples 16700 unjoint_samples 16700 joint_samples 51 [390215, 1022027]
processed_samples 16700 unjoint_samples 16700 joint_samples 49 [1046325, 622870]
processed_samples 16700 unjoint_samples 16700 joint_samples 49 [1046325, 622870]
processed_samples 16700 unjoint_samples 16700 joint_samples 50 [1046975, 760878]
processed_samples 16700 unjoint_samples 16700 joint_samples 50 [1046975, 760878]
processed_samples 16700 unjoint_samples 16700 joint_samples 50 [1046627, 905742]
processed_samples 16700 unjoint_samples 16700 joint_samples 50 [1046627, 905742]
processed_samples 16700 unjoint_samples 16700 joint_samples 49 [822657, 1047364]
processed_samples 16700 unjoint_samples 16700 joint_samples 49 [822657, 1047364]
processed_samples 16700 unjoint_samples 16700 joint_samples 49 [1039244, 1036449]
processed_samples 16700 unjoint_samples 16700 joint_samples 49 [1039244, 1036449]
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a1bb6900] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
processed_samples 16800 unjoint_samples 16800 joint_samples 51 [129900, 1039172]
processed_samples 16800 unjoint_samples 16800 joint_samples 51 [129900, 1039172]
processed_samples 16800 unjoint_samples 16800 joint_samples 50 [1046454, 62458]
processed_samples 16800 unjoint_samples 16800 joint_samples 50 [1046454, 62458]
processed_samples 16800 unjoint_samples 16800 joint_samples 50 [277405, 1047782]
processed_samples 16801 unjoint_samples 16800 joint_samples 51 [280799, 1046595]
processed_samples 16800 unjoint_samples 16800 joint_samples 51 [697675, 1022027]
processed_samples 16800 unjoint_samples 16800 joint_samples 50 [277405, 1047782]
processed_samples 16801 unjoint_samples 16800 joint_samples 51 [1045045, 378671]
processed_samples 16800 unjoint_samples 16800 joint_samples 49 [1046325, 914535]
processed_samples 16800 unjoint_samples 16800 joint_samples 49 [1046325, 914535]
processed_samples 16800 unjoint_samples 16800 joint_samples 51 [697675, 1022027]
processed_samples 16801 unjoint_samples 16800 joint_samples 51 [280799, 1046595]
processed_samples 16801 unjoint_samples 16800 joint_samples 51 [1045045, 378671]
processed_samples 16800 unjoint_samples 16800 joint_samples 50 [1046975, 1016456]
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
processed_samples 16800 unjoint_samples 16800 joint_samples 50 [1046975, 1016456]
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97bab4b80] Missing reference picture, default is 65530
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a22c0d80] Missing reference picture, default is 65530
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
processed_samples 16900 unjoint_samples 16900 joint_samples 51 [1046975, 263040]
processed_samples 16900 unjoint_samples 16900 joint_samples 50 [1046454, 308331]
processed_samples 16900 unjoint_samples 16900 joint_samples 51 [380918, 1039172]
processed_samples 16900 unjoint_samples 16900 joint_samples 50 [326562, 1002060]
processed_samples 16901 unjoint_samples 16900 joint_samples 51 [537580, 1046595]
processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1045045, 703696]
processed_samples 16900 unjoint_samples 16900 joint_samples 51 [950793, 1022027]
processed_samples 16900 unjoint_samples 16900 joint_samples 50 [578464, 1047782]
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
processed_samples 16900 unjoint_samples 16900 joint_samples 51 [1046975, 263040]
processed_samples 16900 unjoint_samples 16900 joint_samples 50 [326562, 1002060]
processed_samples 16900 unjoint_samples 16900 joint_samples 51 [380918, 1039172]
processed_samples 16900 unjoint_samples 16900 joint_samples 50 [1046454, 308331]
processed_samples 16900 unjoint_samples 16900 joint_samples 50 [578464, 1047782]
processed_samples 16900 unjoint_samples 16900 joint_samples 51 [950793, 1022027]
processed_samples 16901 unjoint_samples 16900 joint_samples 51 [537580, 1046595]
processed_samples 16901 unjoint_samples 16900 joint_samples 51 [1045045, 703696]
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97facaf00] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
processed_samples 17000 unjoint_samples 17000 joint_samples 51 [1046975, 586366]
processed_samples 17000 unjoint_samples 17000 joint_samples 50 [636046, 1002060]
processed_samples 17000 unjoint_samples 17000 joint_samples 52 [1046625, 149605]
processed_samples 17000 unjoint_samples 17000 joint_samples 50 [1046454, 600360]
processed_samples 17000 unjoint_samples 17000 joint_samples 51 [669457, 1039172]
processed_samples 17001 unjoint_samples 17000 joint_samples 51 [848640, 1046595]
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1045045, 974136]
processed_samples 17000 unjoint_samples 17000 joint_samples 50 [897129, 1047782]
processed_samples 17000 unjoint_samples 17000 joint_samples 51 [1046975, 586366]
[h264 @ 0x5592a19044c0] mmco: unref short failure
processed_samples 17000 unjoint_samples 17000 joint_samples 52 [1046625, 149605]
processed_samples 17000 unjoint_samples 17000 joint_samples 50 [636046, 1002060]
processed_samples 17000 unjoint_samples 17000 joint_samples 50 [1046454, 600360]
processed_samples 17000 unjoint_samples 17000 joint_samples 51 [669457, 1039172]
processed_samples 17000 unjoint_samples 17000 joint_samples 50 [897129, 1047782]
processed_samples 17001 unjoint_samples 17000 joint_samples 51 [1045045, 974136]
processed_samples 17001 unjoint_samples 17000 joint_samples 51 [848640, 1046595]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 17100 unjoint_samples 17100 joint_samples 52 [1047102, 26597]
processed_samples 17100 unjoint_samples 17100 joint_samples 52 [1047102, 26597]
processed_samples 17100 unjoint_samples 17100 joint_samples 51 [1047431, 95947]
processed_samples 17100 unjoint_samples 17100 joint_samples 51 [1047431, 95947]
processed_samples 17100 unjoint_samples 17100 joint_samples 52 [1046625, 477323]
processed_samples 17100 unjoint_samples 17100 joint_samples 50 [991184, 1002060]
processed_samples 17100 unjoint_samples 17100 joint_samples 52 [1046625, 477323]
processed_samples 17100 unjoint_samples 17100 joint_samples 51 [1046975, 901151]
processed_samples 17100 unjoint_samples 17100 joint_samples 51 [1046975, 901151]
processed_samples 17101 unjoint_samples 17100 joint_samples 52 [279257, 1036324]
processed_samples 17100 unjoint_samples 17100 joint_samples 50 [991184, 1002060]
processed_samples 17101 unjoint_samples 17100 joint_samples 52 [279257, 1036324]
processed_samples 17100 unjoint_samples 17100 joint_samples 50 [1046454, 888060]
processed_samples 17101 unjoint_samples 17100 joint_samples 52 [52896, 1046595]
processed_samples 17101 unjoint_samples 17100 joint_samples 52 [52896, 1046595]
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
processed_samples 17100 unjoint_samples 17100 joint_samples 50 [1046454, 888060]
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97e40bb40] mmco: unref short failure
[h264 @ 0x55d97e40bb40] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97e40bb40] mmco: unref short failure
[h264 @ 0x55d97e40bb40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97e40bb40] mmco: unref short failure
[h264 @ 0x55d97e40bb40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97e40bb40] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97e40bb40] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
processed_samples 17200 unjoint_samples 17200 joint_samples 51 [286437, 1033547]
processed_samples 17200 unjoint_samples 17200 joint_samples 52 [59378, 1046807]
processed_samples 17200 unjoint_samples 17200 joint_samples 52 [1047102, 317709]
processed_samples 17200 unjoint_samples 17200 joint_samples 51 [1046454, 110608]
processed_samples 17200 unjoint_samples 17200 joint_samples 51 [1047431, 378452]
processed_samples 17200 unjoint_samples 17200 joint_samples 51 [286437, 1033547]
processed_samples 17200 unjoint_samples 17200 joint_samples 52 [59378, 1046807]
processed_samples 17201 unjoint_samples 17200 joint_samples 52 [281488, 1046595]
processed_samples 17200 unjoint_samples 17200 joint_samples 52 [1047102, 317709]
processed_samples 17201 unjoint_samples 17200 joint_samples 52 [594143, 1036324]
processed_samples 17200 unjoint_samples 17200 joint_samples 51 [1046454, 110608]
processed_samples 17200 unjoint_samples 17200 joint_samples 52 [1046625, 749159]
processed_samples 17200 unjoint_samples 17200 joint_samples 51 [1047431, 378452]
processed_samples 17201 unjoint_samples 17200 joint_samples 52 [281488, 1046595]
processed_samples 17201 unjoint_samples 17200 joint_samples 52 [594143, 1036324]
[h264 @ 0x55d97dd14f40] mmco: unref short failure
processed_samples 17200 unjoint_samples 17200 joint_samples 52 [1046625, 749159]
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97e729400] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a253a2c0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x55d97e144d00] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
processed_samples 17300 unjoint_samples 17300 joint_samples 53 [44815, 1047288]
processed_samples 17300 unjoint_samples 17300 joint_samples 52 [338287, 1046807]
processed_samples 17300 unjoint_samples 17300 joint_samples 51 [1046454, 481139]
processed_samples 17300 unjoint_samples 17300 joint_samples 51 [510665, 1033547]
processed_samples 17300 unjoint_samples 17300 joint_samples 51 [1047431, 652299]
processed_samples 17300 unjoint_samples 17300 joint_samples 52 [1047102, 604950]
processed_samples 17301 unjoint_samples 17300 joint_samples 52 [693756, 1046595]
processed_samples 17301 unjoint_samples 17300 joint_samples 52 [916233, 1036324]
processed_samples 17300 unjoint_samples 17300 joint_samples 51 [1046454, 481139]
processed_samples 17300 unjoint_samples 17300 joint_samples 53 [44815, 1047288]
processed_samples 17300 unjoint_samples 17300 joint_samples 52 [338287, 1046807]
processed_samples 17300 unjoint_samples 17300 joint_samples 51 [510665, 1033547]
processed_samples 17301 unjoint_samples 17300 joint_samples 52 [693756, 1046595]
processed_samples 17300 unjoint_samples 17300 joint_samples 52 [1047102, 604950]
processed_samples 17300 unjoint_samples 17300 joint_samples 51 [1047431, 652299]
processed_samples 17301 unjoint_samples 17300 joint_samples 52 [916233, 1036324]
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
processed_samples 17400 unjoint_samples 17400 joint_samples 53 [359418, 1047288]
processed_samples 17400 unjoint_samples 17400 joint_samples 52 [699690, 1046807]
processed_samples 17400 unjoint_samples 17400 joint_samples 52 [1047102, 922680]
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 17400 unjoint_samples 17400 joint_samples 51 [764901, 1033547]
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
processed_samples 17400 unjoint_samples 17400 joint_samples 51 [1046454, 767990]
processed_samples 17400 unjoint_samples 17400 joint_samples 51 [1047431, 1007326]
processed_samples 17401 unjoint_samples 17400 joint_samples 53 [125744, 1040579]
processed_samples 17401 unjoint_samples 17400 joint_samples 52 [980494, 1046595]
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
processed_samples 17400 unjoint_samples 17400 joint_samples 53 [359418, 1047288]
processed_samples 17400 unjoint_samples 17400 joint_samples 52 [699690, 1046807]
processed_samples 17400 unjoint_samples 17400 joint_samples 52 [1047102, 922680]
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
processed_samples 17400 unjoint_samples 17400 joint_samples 51 [1046454, 767990]
processed_samples 17400 unjoint_samples 17400 joint_samples 51 [764901, 1033547]
processed_samples 17401 unjoint_samples 17400 joint_samples 53 [125744, 1040579]
processed_samples 17400 unjoint_samples 17400 joint_samples 51 [1047431, 1007326]
processed_samples 17401 unjoint_samples 17400 joint_samples 52 [980494, 1046595]
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a73af580] mmco: unref short failure
[h264 @ 0x5592a73af580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97a5beec0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a045efc0] mmco: unref short failure
[h264 @ 0x5592a045efc0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a4e3aa40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
processed_samples 17500 unjoint_samples 17500 joint_samples 53 [626776, 1047288]
processed_samples 17500 unjoint_samples 17500 joint_samples 53 [243477, 1038729]
processed_samples 17500 unjoint_samples 17500 joint_samples 52 [1022021, 92756]
processed_samples 17500 unjoint_samples 17500 joint_samples 52 [261436, 1047292]
processed_samples 17501 unjoint_samples 17500 joint_samples 53 [434081, 1040579]
[h264 @ 0x5592a529c8c0] mmco: unref short failure
processed_samples 17501 unjoint_samples 17500 joint_samples 53 [1019152, 385563]
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 17500 unjoint_samples 17500 joint_samples 52 [930155, 1046807]
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 17500 unjoint_samples 17500 joint_samples 51 [1046454, 995135]
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
processed_samples 17500 unjoint_samples 17500 joint_samples 53 [243477, 1038729]
processed_samples 17500 unjoint_samples 17500 joint_samples 53 [626776, 1047288]
processed_samples 17500 unjoint_samples 17500 joint_samples 52 [1022021, 92756]
processed_samples 17500 unjoint_samples 17500 joint_samples 52 [930155, 1046807]
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
processed_samples 17500 unjoint_samples 17500 joint_samples 52 [261436, 1047292]
processed_samples 17501 unjoint_samples 17500 joint_samples 53 [434081, 1040579]
[h264 @ 0x55d97bab4b80] mmco: unref short failure
processed_samples 17501 unjoint_samples 17500 joint_samples 53 [1019152, 385563]
processed_samples 17500 unjoint_samples 17500 joint_samples 51 [1046454, 995135]
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
processed_samples 17600 unjoint_samples 17600 joint_samples 52 [353712, 1024452]
processed_samples 17600 unjoint_samples 17600 joint_samples 52 [1022021, 353288]
processed_samples 17600 unjoint_samples 17600 joint_samples 53 [1044335, 149815]
processed_samples 17600 unjoint_samples 17600 joint_samples 53 [495346, 1038729]
processed_samples 17601 unjoint_samples 17600 joint_samples 53 [1019152, 733682]
processed_samples 17600 unjoint_samples 17600 joint_samples 53 [934677, 1047288]
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
processed_samples 17600 unjoint_samples 17600 joint_samples 52 [596292, 1047292]
processed_samples 17601 unjoint_samples 17600 joint_samples 53 [719667, 1040579]
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
processed_samples 17600 unjoint_samples 17600 joint_samples 53 [1044335, 149815]
processed_samples 17600 unjoint_samples 17600 joint_samples 52 [1022021, 353288]
processed_samples 17600 unjoint_samples 17600 joint_samples 53 [495346, 1038729]
processed_samples 17600 unjoint_samples 17600 joint_samples 52 [353712, 1024452]
processed_samples 17600 unjoint_samples 17600 joint_samples 52 [596292, 1047292]
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 17601 unjoint_samples 17600 joint_samples 53 [1019152, 733682]
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 17600 unjoint_samples 17600 joint_samples 53 [934677, 1047288]
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 17601 unjoint_samples 17600 joint_samples 53 [719667, 1040579]
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 17700 unjoint_samples 17700 joint_samples 52 [666829, 1024452]
processed_samples 17700 unjoint_samples 17700 joint_samples 54 [1045170, 103816]
processed_samples 17700 unjoint_samples 17700 joint_samples 53 [1044335, 424998]
processed_samples 17700 unjoint_samples 17700 joint_samples 53 [732325, 1038729]
processed_samples 17700 unjoint_samples 17700 joint_samples 52 [1022021, 700161]
processed_samples 17701 unjoint_samples 17700 joint_samples 53 [1027170, 1029839]
[h264 @ 0x5592a0d96180] mmco: unref short failure
processed_samples 17700 unjoint_samples 17700 joint_samples 52 [824617, 1047292]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 17701 unjoint_samples 17700 joint_samples 53 [1019565, 1040579]
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
processed_samples 17700 unjoint_samples 17700 joint_samples 52 [666829, 1024452]
processed_samples 17700 unjoint_samples 17700 joint_samples 54 [1045170, 103816]
processed_samples 17700 unjoint_samples 17700 joint_samples 53 [1044335, 424998]
processed_samples 17700 unjoint_samples 17700 joint_samples 52 [1022021, 700161]
processed_samples 17700 unjoint_samples 17700 joint_samples 53 [732325, 1038729]
processed_samples 17701 unjoint_samples 17700 joint_samples 53 [1027170, 1029839]
processed_samples 17701 unjoint_samples 17700 joint_samples 53 [1019565, 1040579]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 17700 unjoint_samples 17700 joint_samples 52 [824617, 1047292]
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1012051, 161252]
processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1045170, 538131]
processed_samples 17800 unjoint_samples 17800 joint_samples 53 [116142, 1047292]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 17800 unjoint_samples 17800 joint_samples 52 [1022021, 988291]
processed_samples 17800 unjoint_samples 17800 joint_samples 53 [1044335, 811764]
processed_samples 17801 unjoint_samples 17800 joint_samples 54 [203655, 1046765]
processed_samples 17801 unjoint_samples 17800 joint_samples 54 [238021, 1042770]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1012051, 161252]
processed_samples 17800 unjoint_samples 17800 joint_samples 52 [965324, 1024452]
processed_samples 17800 unjoint_samples 17800 joint_samples 53 [116142, 1047292]
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 17800 unjoint_samples 17800 joint_samples 54 [1045170, 538131]
processed_samples 17801 unjoint_samples 17800 joint_samples 54 [203655, 1046765]
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
processed_samples 17801 unjoint_samples 17800 joint_samples 54 [238021, 1042770]
processed_samples 17800 unjoint_samples 17800 joint_samples 52 [965324, 1024452]
processed_samples 17800 unjoint_samples 17800 joint_samples 53 [1044335, 811764]
processed_samples 17800 unjoint_samples 17800 joint_samples 52 [1022021, 988291]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
processed_samples 17900 unjoint_samples 17900 joint_samples 54 [80545, 1044281]
processed_samples 17900 unjoint_samples 17900 joint_samples 54 [1012051, 422366]
processed_samples 17900 unjoint_samples 17900 joint_samples 53 [251267, 1047409]
processed_samples 17900 unjoint_samples 17900 joint_samples 54 [80545, 1044281]
processed_samples 17901 unjoint_samples 17900 joint_samples 54 [624869, 1046765]
processed_samples 17900 unjoint_samples 17900 joint_samples 54 [1012051, 422366]
processed_samples 17900 unjoint_samples 17900 joint_samples 53 [251267, 1047409]
processed_samples 17900 unjoint_samples 17900 joint_samples 53 [518511, 1047292]
processed_samples 17900 unjoint_samples 17900 joint_samples 53 [289375, 1034206]
processed_samples 17901 unjoint_samples 17900 joint_samples 54 [624869, 1046765]
processed_samples 17901 unjoint_samples 17900 joint_samples 54 [560294, 1042770]
processed_samples 17900 unjoint_samples 17900 joint_samples 53 [518511, 1047292]
processed_samples 17900 unjoint_samples 17900 joint_samples 54 [1045170, 779977]
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
processed_samples 17900 unjoint_samples 17900 joint_samples 53 [289375, 1034206]
processed_samples 17900 unjoint_samples 17900 joint_samples 54 [1045170, 779977]
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
processed_samples 17901 unjoint_samples 17900 joint_samples 54 [560294, 1042770]
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a1fa3f40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a4e9a600] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
processed_samples 18000 unjoint_samples 18000 joint_samples 55 [1047608, 10366]
processed_samples 18000 unjoint_samples 18000 joint_samples 55 [1047608, 10366]
processed_samples 18000 unjoint_samples 18000 joint_samples 54 [1012051, 698793]
processed_samples 18000 unjoint_samples 18000 joint_samples 54 [1012051, 698793]
processed_samples 18000 unjoint_samples 18000 joint_samples 53 [508438, 1047409]
processed_samples 18000 unjoint_samples 18000 joint_samples 53 [508438, 1047409]
processed_samples 18000 unjoint_samples 18000 joint_samples 54 [380022, 1044281]
processed_samples 18000 unjoint_samples 18000 joint_samples 54 [380022, 1044281]
processed_samples 18000 unjoint_samples 18000 joint_samples 53 [761796, 1047292]
processed_samples 18000 unjoint_samples 18000 joint_samples 53 [761796, 1047292]
processed_samples 18000 unjoint_samples 18000 joint_samples 53 [638855, 1034206]
processed_samples 18001 unjoint_samples 18000 joint_samples 54 [961504, 1046765]
processed_samples 18001 unjoint_samples 18000 joint_samples 54 [961504, 1046765]
processed_samples 18001 unjoint_samples 18000 joint_samples 54 [811037, 1042770]
processed_samples 18000 unjoint_samples 18000 joint_samples 53 [638855, 1034206]
processed_samples 18001 unjoint_samples 18000 joint_samples 54 [811037, 1042770]
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 18100 unjoint_samples 18100 joint_samples 55 [1047608, 334516]
processed_samples 18100 unjoint_samples 18100 joint_samples 54 [649553, 1044281]
processed_samples 18100 unjoint_samples 18100 joint_samples 54 [1027273, 1026459]
processed_samples 18100 unjoint_samples 18100 joint_samples 54 [1045710, 127282]
processed_samples 18101 unjoint_samples 18100 joint_samples 55 [1045982, 112147]
processed_samples 18101 unjoint_samples 18100 joint_samples 55 [160281, 1046765]
processed_samples 18100 unjoint_samples 18100 joint_samples 53 [871744, 1047409]
processed_samples 18100 unjoint_samples 18100 joint_samples 53 [946758, 1034206]
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
processed_samples 18100 unjoint_samples 18100 joint_samples 55 [1047608, 334516]
processed_samples 18100 unjoint_samples 18100 joint_samples 54 [1045710, 127282]
processed_samples 18100 unjoint_samples 18100 joint_samples 53 [946758, 1034206]
processed_samples 18100 unjoint_samples 18100 joint_samples 54 [649553, 1044281]
processed_samples 18100 unjoint_samples 18100 joint_samples 53 [871744, 1047409]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 18100 unjoint_samples 18100 joint_samples 54 [1027273, 1026459]
processed_samples 18101 unjoint_samples 18100 joint_samples 55 [1045982, 112147]
processed_samples 18101 unjoint_samples 18100 joint_samples 55 [160281, 1046765]
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d983320600] mmco: unref short failure
[h264 @ 0x55d983320600] mmco: unref short failure
[h264 @ 0x55d983320600] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a10d6340] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
processed_samples 18200 unjoint_samples 18200 joint_samples 54 [191925, 1047409]
processed_samples 18200 unjoint_samples 18200 joint_samples 55 [446984, 1032412]
processed_samples 18200 unjoint_samples 18200 joint_samples 54 [1047370, 154196]
processed_samples 18200 unjoint_samples 18200 joint_samples 54 [1045710, 434693]
processed_samples 18200 unjoint_samples 18200 joint_samples 54 [845187, 1044281]
processed_samples 18201 unjoint_samples 18200 joint_samples 55 [511431, 1046765]
processed_samples 18201 unjoint_samples 18200 joint_samples 55 [1045982, 531686]
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
processed_samples 18200 unjoint_samples 18200 joint_samples 55 [1047608, 638976]
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
processed_samples 18200 unjoint_samples 18200 joint_samples 54 [191925, 1047409]
processed_samples 18200 unjoint_samples 18200 joint_samples 55 [446984, 1032412]
processed_samples 18200 unjoint_samples 18200 joint_samples 54 [1047370, 154196]
processed_samples 18200 unjoint_samples 18200 joint_samples 54 [1045710, 434693]
processed_samples 18201 unjoint_samples 18200 joint_samples 55 [511431, 1046765]
processed_samples 18200 unjoint_samples 18200 joint_samples 54 [845187, 1044281]
processed_samples 18201 unjoint_samples 18200 joint_samples 55 [1045982, 531686]
processed_samples 18200 unjoint_samples 18200 joint_samples 55 [1047608, 638976]
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a18dbc80] mmco: unref short failure
[h264 @ 0x5592a18dbc80] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a618b9c0] mmco: unref short failure
processed_samples 18300 unjoint_samples 18300 joint_samples 55 [58624, 1046535]
processed_samples 18300 unjoint_samples 18300 joint_samples 54 [1047370, 440929]
processed_samples 18300 unjoint_samples 18300 joint_samples 55 [861529, 1032412]
processed_samples 18300 unjoint_samples 18300 joint_samples 54 [1045710, 714034]
processed_samples 18300 unjoint_samples 18300 joint_samples 54 [622769, 1047409]
processed_samples 18300 unjoint_samples 18300 joint_samples 55 [1047608, 864600]
processed_samples 18301 unjoint_samples 18300 joint_samples 55 [836451, 1046765]
processed_samples 18301 unjoint_samples 18300 joint_samples 55 [1045982, 800675]
processed_samples 18300 unjoint_samples 18300 joint_samples 55 [861529, 1032412]
processed_samples 18300 unjoint_samples 18300 joint_samples 55 [58624, 1046535]
processed_samples 18300 unjoint_samples 18300 joint_samples 55 [1047608, 864600]
processed_samples 18300 unjoint_samples 18300 joint_samples 54 [1047370, 440929]
processed_samples 18300 unjoint_samples 18300 joint_samples 54 [622769, 1047409]
processed_samples 18300 unjoint_samples 18300 joint_samples 54 [1045710, 714034]
processed_samples 18301 unjoint_samples 18300 joint_samples 55 [836451, 1046765]
processed_samples 18301 unjoint_samples 18300 joint_samples 55 [1045982, 800675]
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a479cd80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x5592a105f780] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d983320600] mmco: unref short failure
[h264 @ 0x55d983320600] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
processed_samples 18400 unjoint_samples 18400 joint_samples 56 [1031722, 88312]
processed_samples 18400 unjoint_samples 18400 joint_samples 56 [1047608, 79306]
processed_samples 18400 unjoint_samples 18400 joint_samples 55 [315380, 1046535]
processed_samples 18400 unjoint_samples 18400 joint_samples 54 [1047370, 644523]
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
processed_samples 18401 unjoint_samples 18400 joint_samples 56 [46914, 1036795]
processed_samples 18401 unjoint_samples 18400 joint_samples 56 [1031658, 150490]
processed_samples 18400 unjoint_samples 18400 joint_samples 54 [906558, 1047409]
processed_samples 18400 unjoint_samples 18400 joint_samples 54 [1045710, 1042275]
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97adfba00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
processed_samples 18400 unjoint_samples 18400 joint_samples 56 [1031722, 88312]
processed_samples 18400 unjoint_samples 18400 joint_samples 56 [1047608, 79306]
processed_samples 18400 unjoint_samples 18400 joint_samples 55 [315380, 1046535]
processed_samples 18400 unjoint_samples 18400 joint_samples 54 [1047370, 644523]
processed_samples 18400 unjoint_samples 18400 joint_samples 54 [906558, 1047409]
[h264 @ 0x5592a489d240] mmco: unref short failure
processed_samples 18401 unjoint_samples 18400 joint_samples 56 [46914, 1036795]
processed_samples 18401 unjoint_samples 18400 joint_samples 56 [1031658, 150490]
processed_samples 18400 unjoint_samples 18400 joint_samples 54 [1045710, 1042275]
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d983320600] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a4cbed40] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a9a16b00] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
processed_samples 18500 unjoint_samples 18500 joint_samples 55 [1047378, 229474]
processed_samples 18500 unjoint_samples 18500 joint_samples 56 [1047608, 405850]
[h264 @ 0x5592a9a59e00] mmco: unref short failure
processed_samples 18500 unjoint_samples 18500 joint_samples 55 [1047378, 229474]
processed_samples 18500 unjoint_samples 18500 joint_samples 56 [1031722, 349094]
[h264 @ 0x5592a23c6a40] mmco: unref short failure
processed_samples 18500 unjoint_samples 18500 joint_samples 56 [1047608, 405850]
processed_samples 18500 unjoint_samples 18500 joint_samples 55 [596035, 1046535]
processed_samples 18500 unjoint_samples 18500 joint_samples 56 [1031722, 349094]
processed_samples 18500 unjoint_samples 18500 joint_samples 54 [1047370, 905044]
processed_samples 18500 unjoint_samples 18500 joint_samples 55 [245936, 1046804]
processed_samples 18500 unjoint_samples 18500 joint_samples 54 [1047370, 905044]
processed_samples 18501 unjoint_samples 18500 joint_samples 56 [466664, 1036795]
[h264 @ 0x55d97b455300] mmco: unref short failure
processed_samples 18500 unjoint_samples 18500 joint_samples 55 [245936, 1046804]
[h264 @ 0x55d982fa7a40] mmco: unref short failure
processed_samples 18501 unjoint_samples 18500 joint_samples 56 [1031658, 434214]
processed_samples 18501 unjoint_samples 18500 joint_samples 56 [466664, 1036795]
processed_samples 18500 unjoint_samples 18500 joint_samples 55 [596035, 1046535]
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
processed_samples 18501 unjoint_samples 18500 joint_samples 56 [1031658, 434214]
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97a92a340] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97b969680] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x5592a28d0000] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a0f16a80] mmco: unref short failure
[h264 @ 0x5592a0f16a80] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
processed_samples 18600 unjoint_samples 18600 joint_samples 55 [999784, 1046535]
processed_samples 18600 unjoint_samples 18600 joint_samples 55 [128654, 1047413]
processed_samples 18600 unjoint_samples 18600 joint_samples 56 [1031722, 590547]
processed_samples 18600 unjoint_samples 18600 joint_samples 56 [1047608, 648414]
processed_samples 18600 unjoint_samples 18600 joint_samples 55 [1047378, 600505]
processed_samples 18600 unjoint_samples 18600 joint_samples 55 [611787, 1046804]
processed_samples 18601 unjoint_samples 18600 joint_samples 56 [737826, 1036795]
processed_samples 18601 unjoint_samples 18600 joint_samples 56 [1031658, 761620]
processed_samples 18600 unjoint_samples 18600 joint_samples 56 [1031722, 590547]
processed_samples 18600 unjoint_samples 18600 joint_samples 55 [1047378, 600505]
processed_samples 18600 unjoint_samples 18600 joint_samples 55 [128654, 1047413]
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
processed_samples 18600 unjoint_samples 18600 joint_samples 55 [611787, 1046804]
processed_samples 18600 unjoint_samples 18600 joint_samples 56 [1047608, 648414]
processed_samples 18601 unjoint_samples 18600 joint_samples 56 [1031658, 761620]
processed_samples 18600 unjoint_samples 18600 joint_samples 55 [999784, 1046535]
processed_samples 18601 unjoint_samples 18600 joint_samples 56 [737826, 1036795]
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x5592a4ee5d40] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b98b840] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1047608, 946069]
processed_samples 18700 unjoint_samples 18700 joint_samples 56 [370383, 1046535]
processed_samples 18700 unjoint_samples 18700 joint_samples 55 [549525, 1047413]
processed_samples 18700 unjoint_samples 18700 joint_samples 55 [1047378, 932681]
processed_samples 18701 unjoint_samples 18700 joint_samples 56 [1033665, 1033744]
processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1031722, 900262]
processed_samples 18700 unjoint_samples 18700 joint_samples 55 [847881, 1046804]
processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1047608, 946069]
processed_samples 18700 unjoint_samples 18700 joint_samples 56 [370383, 1046535]
processed_samples 18701 unjoint_samples 18700 joint_samples 57 [1023679, 49766]
processed_samples 18700 unjoint_samples 18700 joint_samples 55 [549525, 1047413]
processed_samples 18700 unjoint_samples 18700 joint_samples 55 [1047378, 932681]
processed_samples 18700 unjoint_samples 18700 joint_samples 56 [1031722, 900262]
processed_samples 18701 unjoint_samples 18700 joint_samples 56 [1033665, 1033744]
processed_samples 18700 unjoint_samples 18700 joint_samples 55 [847881, 1046804]
processed_samples 18701 unjoint_samples 18700 joint_samples 57 [1023679, 49766]
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x55d97ac563c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a462ea80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x55d97cd63980] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a1ec92c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
processed_samples 18800 unjoint_samples 18800 joint_samples 57 [81153, 1048155]
processed_samples 18800 unjoint_samples 18800 joint_samples 57 [81153, 1048155]
processed_samples 18800 unjoint_samples 18800 joint_samples 56 [777142, 1046535]
processed_samples 18800 unjoint_samples 18800 joint_samples 57 [289733, 1035926]
processed_samples 18800 unjoint_samples 18800 joint_samples 57 [289733, 1035926]
processed_samples 18800 unjoint_samples 18800 joint_samples 56 [418973, 1045923]
processed_samples 18800 unjoint_samples 18800 joint_samples 56 [777142, 1046535]
processed_samples 18800 unjoint_samples 18800 joint_samples 56 [418973, 1045923]
processed_samples 18800 unjoint_samples 18800 joint_samples 56 [87141, 1047742]
processed_samples 18800 unjoint_samples 18800 joint_samples 56 [87141, 1047742]
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
processed_samples 18801 unjoint_samples 18800 joint_samples 57 [1033665, 350743]
processed_samples 18801 unjoint_samples 18800 joint_samples 57 [1023679, 371183]
processed_samples 18801 unjoint_samples 18800 joint_samples 57 [1033665, 350743]
processed_samples 18801 unjoint_samples 18800 joint_samples 57 [1023679, 371183]
processed_samples 18800 unjoint_samples 18800 joint_samples 55 [908846, 1047413]
processed_samples 18800 unjoint_samples 18800 joint_samples 55 [908846, 1047413]
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b32cf80] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a159cec0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d97df82080] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
processed_samples 18900 unjoint_samples 18900 joint_samples 57 [74222, 1046535]
processed_samples 18900 unjoint_samples 18900 joint_samples 57 [303643, 1048155]
processed_samples 18900 unjoint_samples 18900 joint_samples 56 [997473, 285327]
processed_samples 18900 unjoint_samples 18900 joint_samples 56 [719944, 1045923]
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
processed_samples 18900 unjoint_samples 18900 joint_samples 57 [593972, 1035926]
[h264 @ 0x55d97ee241c0] mmco: unref short failure
processed_samples 18900 unjoint_samples 18900 joint_samples 56 [431480, 1047742]
processed_samples 18900 unjoint_samples 18900 joint_samples 56 [997473, 285327]
processed_samples 18900 unjoint_samples 18900 joint_samples 57 [74222, 1046535]
processed_samples 18900 unjoint_samples 18900 joint_samples 57 [303643, 1048155]
processed_samples 18901 unjoint_samples 18900 joint_samples 57 [1023679, 836633]
processed_samples 18901 unjoint_samples 18900 joint_samples 57 [1033665, 652537]
processed_samples 18900 unjoint_samples 18900 joint_samples 56 [719944, 1045923]
processed_samples 18900 unjoint_samples 18900 joint_samples 56 [431480, 1047742]
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
processed_samples 18900 unjoint_samples 18900 joint_samples 57 [593972, 1035926]
processed_samples 18901 unjoint_samples 18900 joint_samples 57 [1023679, 836633]
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
processed_samples 18901 unjoint_samples 18900 joint_samples 57 [1033665, 652537]
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x5592a537b180] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
processed_samples 19000 unjoint_samples 19000 joint_samples 57 [685059, 1048155]
processed_samples 19000 unjoint_samples 19000 joint_samples 56 [997473, 485703]
processed_samples 19000 unjoint_samples 19000 joint_samples 56 [1012613, 1045923]
processed_samples 19000 unjoint_samples 19000 joint_samples 56 [1012613, 1045923]
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
processed_samples 19000 unjoint_samples 19000 joint_samples 56 [736897, 1047742]
processed_samples 19001 unjoint_samples 19000 joint_samples 58 [1042013, 48740]
processed_samples 19000 unjoint_samples 19000 joint_samples 57 [388801, 1046535]
processed_samples 19000 unjoint_samples 19000 joint_samples 56 [997473, 485703]
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
processed_samples 19000 unjoint_samples 19000 joint_samples 57 [388801, 1046535]
processed_samples 19000 unjoint_samples 19000 joint_samples 57 [834883, 1035926]
processed_samples 19001 unjoint_samples 19000 joint_samples 58 [1042013, 48740]
processed_samples 19000 unjoint_samples 19000 joint_samples 57 [685059, 1048155]
processed_samples 19000 unjoint_samples 19000 joint_samples 56 [736897, 1047742]
processed_samples 19000 unjoint_samples 19000 joint_samples 57 [834883, 1035926]
processed_samples 19001 unjoint_samples 19000 joint_samples 57 [1033665, 1011449]
processed_samples 19001 unjoint_samples 19000 joint_samples 57 [1033665, 1011449]
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a48aaa80] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b79e140] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x55d97b701c40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d981fe26c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a2abfe80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x5592a550a740] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97be2e400] mmco: unref short failure
[h264 @ 0x55d97be9d240] Missing reference picture, default is 65530
[h264 @ 0x55d97be9d240] Missing reference picture, default is 65530
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x5592a4deb500] Missing reference picture, default is 65530
[h264 @ 0x5592a4deb500] Missing reference picture, default is 65530
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
processed_samples 19100 unjoint_samples 19100 joint_samples 58 [89346, 1045000]
processed_samples 19100 unjoint_samples 19100 joint_samples 57 [1044153, 24487]
processed_samples 19100 unjoint_samples 19100 joint_samples 57 [1030768, 287055]
processed_samples 19100 unjoint_samples 19100 joint_samples 56 [997473, 817479]
processed_samples 19100 unjoint_samples 19100 joint_samples 58 [89346, 1045000]
processed_samples 19100 unjoint_samples 19100 joint_samples 57 [1044153, 24487]
processed_samples 19100 unjoint_samples 19100 joint_samples 57 [1030768, 287055]
processed_samples 19100 unjoint_samples 19100 joint_samples 57 [708395, 1046535]
processed_samples 19101 unjoint_samples 19100 joint_samples 58 [388937, 1013574]
processed_samples 19100 unjoint_samples 19100 joint_samples 56 [997473, 817479]
processed_samples 19101 unjoint_samples 19100 joint_samples 58 [1042013, 320248]
processed_samples 19101 unjoint_samples 19100 joint_samples 58 [388937, 1013574]
processed_samples 19101 unjoint_samples 19100 joint_samples 58 [1042013, 320248]
processed_samples 19100 unjoint_samples 19100 joint_samples 57 [708395, 1046535]
processed_samples 19100 unjoint_samples 19100 joint_samples 58 [3399, 1048155]
processed_samples 19100 unjoint_samples 19100 joint_samples 58 [3399, 1048155]
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d97ac78600] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97f7d2640] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x55d97a81b380] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a13cdd80] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 19200 unjoint_samples 19200 joint_samples 58 [259853, 1048155]
processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1030768, 641013]
processed_samples 19200 unjoint_samples 19200 joint_samples 58 [430831, 1045000]
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1044153, 450388]
processed_samples 19201 unjoint_samples 19200 joint_samples 58 [1042013, 629137]
processed_samples 19201 unjoint_samples 19200 joint_samples 58 [705339, 1013574]
processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1002082, 1046535]
processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1047231, 2112]
processed_samples 19200 unjoint_samples 19200 joint_samples 58 [259853, 1048155]
processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1030768, 641013]
processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1044153, 450388]
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
processed_samples 19200 unjoint_samples 19200 joint_samples 58 [430831, 1045000]
processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1047231, 2112]
processed_samples 19201 unjoint_samples 19200 joint_samples 58 [705339, 1013574]
processed_samples 19201 unjoint_samples 19200 joint_samples 58 [1042013, 629137]
processed_samples 19200 unjoint_samples 19200 joint_samples 57 [1002082, 1046535]
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x55d97e68d780] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b77a4c0] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e545600] mmco: unref short failure
[h264 @ 0x55d97e545600] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x55d97b6c8400] mmco: unref short failure
processed_samples 19300 unjoint_samples 19300 joint_samples 58 [536986, 1048155]
processed_samples 19300 unjoint_samples 19300 joint_samples 58 [730732, 1045000]
processed_samples 19300 unjoint_samples 19300 joint_samples 58 [191349, 1046535]
processed_samples 19300 unjoint_samples 19300 joint_samples 57 [1047231, 416958]
processed_samples 19300 unjoint_samples 19300 joint_samples 57 [1044153, 718068]
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
processed_samples 19301 unjoint_samples 19300 joint_samples 58 [1042013, 856023]
processed_samples 19300 unjoint_samples 19300 joint_samples 57 [1030768, 987233]
processed_samples 19300 unjoint_samples 19300 joint_samples 58 [536986, 1048155]
processed_samples 19301 unjoint_samples 19300 joint_samples 58 [1009052, 1013574]
processed_samples 19300 unjoint_samples 19300 joint_samples 58 [730732, 1045000]
processed_samples 19300 unjoint_samples 19300 joint_samples 58 [191349, 1046535]
processed_samples 19300 unjoint_samples 19300 joint_samples 57 [1047231, 416958]
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
processed_samples 19300 unjoint_samples 19300 joint_samples 57 [1044153, 718068]
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
processed_samples 19300 unjoint_samples 19300 joint_samples 57 [1030768, 987233]
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
processed_samples 19301 unjoint_samples 19300 joint_samples 58 [1042013, 856023]
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
processed_samples 19301 unjoint_samples 19300 joint_samples 58 [1009052, 1013574]
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97b442840] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [1026197, 1045000]
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [441037, 1046535]
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [60808, 1039287]
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [306099, 1022392]
processed_samples 19401 unjoint_samples 19400 joint_samples 59 [110742, 1046081]
[h264 @ 0x55d9819618c0] mmco: unref short failure
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [876829, 1048155]
processed_samples 19401 unjoint_samples 19400 joint_samples 59 [468614, 1020366]
processed_samples 19400 unjoint_samples 19400 joint_samples 57 [1047231, 687912]
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97ae1a9c0] mmco: unref short failure
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [1026197, 1045000]
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [441037, 1046535]
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [60808, 1039287]
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [306099, 1022392]
processed_samples 19400 unjoint_samples 19400 joint_samples 57 [1047231, 687912]
processed_samples 19401 unjoint_samples 19400 joint_samples 59 [110742, 1046081]
[h264 @ 0x5592a13cdd80] mmco: unref short failure
processed_samples 19401 unjoint_samples 19400 joint_samples 59 [468614, 1020366]
processed_samples 19400 unjoint_samples 19400 joint_samples 58 [876829, 1048155]
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x5592a18f7b00] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x55d97adf0480] mmco: unref short failure
[h264 @ 0x5592a106e800] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97e210840] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
processed_samples 19500 unjoint_samples 19500 joint_samples 59 [1036772, 148223]
processed_samples 19500 unjoint_samples 19500 joint_samples 59 [254127, 1046332]
processed_samples 19500 unjoint_samples 19500 joint_samples 58 [331924, 1039287]
processed_samples 19500 unjoint_samples 19500 joint_samples 57 [1047231, 993025]
processed_samples 19500 unjoint_samples 19500 joint_samples 58 [608877, 1022392]
processed_samples 19501 unjoint_samples 19500 joint_samples 59 [345554, 1046081]
processed_samples 19501 unjoint_samples 19500 joint_samples 59 [689179, 1020366]
processed_samples 19500 unjoint_samples 19500 joint_samples 58 [747699, 1046535]
processed_samples 19500 unjoint_samples 19500 joint_samples 59 [254127, 1046332]
processed_samples 19500 unjoint_samples 19500 joint_samples 59 [1036772, 148223]
processed_samples 19500 unjoint_samples 19500 joint_samples 58 [608877, 1022392]
processed_samples 19500 unjoint_samples 19500 joint_samples 57 [1047231, 993025]
processed_samples 19500 unjoint_samples 19500 joint_samples 58 [747699, 1046535]
processed_samples 19500 unjoint_samples 19500 joint_samples 58 [331924, 1039287]
processed_samples 19501 unjoint_samples 19500 joint_samples 59 [345554, 1046081]
processed_samples 19501 unjoint_samples 19500 joint_samples 59 [689179, 1020366]
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x55d97ad69a80] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a46eec00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a7b17200] mmco: unref short failure
[h264 @ 0x5592a7b17200] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x5592a97c0940] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x55d97c2b5880] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
processed_samples 19600 unjoint_samples 19600 joint_samples 59 [1036772, 446406]
processed_samples 19600 unjoint_samples 19600 joint_samples 58 [860983, 1022392]
processed_samples 19600 unjoint_samples 19600 joint_samples 59 [650058, 1046332]
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
processed_samples 19600 unjoint_samples 19600 joint_samples 58 [1047231, 176934]
processed_samples 19600 unjoint_samples 19600 joint_samples 58 [1034473, 1046535]
processed_samples 19600 unjoint_samples 19600 joint_samples 58 [610174, 1039287]
processed_samples 19601 unjoint_samples 19600 joint_samples 59 [936087, 1020366]
processed_samples 19601 unjoint_samples 19600 joint_samples 59 [671038, 1046081]
processed_samples 19600 unjoint_samples 19600 joint_samples 59 [1036772, 446406]
processed_samples 19600 unjoint_samples 19600 joint_samples 58 [860983, 1022392]
processed_samples 19600 unjoint_samples 19600 joint_samples 59 [650058, 1046332]
processed_samples 19600 unjoint_samples 19600 joint_samples 58 [1047231, 176934]
processed_samples 19600 unjoint_samples 19600 joint_samples 58 [610174, 1039287]
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
processed_samples 19601 unjoint_samples 19600 joint_samples 59 [671038, 1046081]
processed_samples 19601 unjoint_samples 19600 joint_samples 59 [936087, 1020366]
processed_samples 19600 unjoint_samples 19600 joint_samples 58 [1034473, 1046535]
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x5592a48b10c0] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x55d97e1d0540] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97ad8efc0] mmco: unref short failure
[h264 @ 0x55d97ad8efc0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x5592a140f7c0] mmco: unref short failure
[h264 @ 0x55d97e3ea240] illegal short term buffer state detected
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a48560c0] illegal short term buffer state detected
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x55d980e9c140] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x55d97c0401c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x5592a1c80f80] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x55d979c3b540] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a18f7b00] [h264 @ 0x55d97e210840] mmco: unref short failure
mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a109e4c0] mmco: unref short failure
[h264 @ 0x55d97e79fb80] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a0d97bc0] mmco: unref short failure
[h264 @ 0x5592a0d97bc0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1047972, 314370]
processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1036772, 663728]
processed_samples 19700 unjoint_samples 19700 joint_samples 59 [233679, 1047121]
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
processed_samples 19700 unjoint_samples 19700 joint_samples 58 [1047231, 447843]
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
processed_samples 19700 unjoint_samples 19700 joint_samples 58 [872303, 1039287]
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
processed_samples 19701 unjoint_samples 19700 joint_samples 60 [185877, 1046295]
processed_samples 19700 unjoint_samples 19700 joint_samples 59 [930497, 1046332]
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 19700 unjoint_samples 19700 joint_samples 58 [1047231, 447843]
processed_samples 19701 unjoint_samples 19700 joint_samples 59 [1001129, 1046081]
processed_samples 19700 unjoint_samples 19700 joint_samples 59 [233679, 1047121]
processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1047972, 314370]
processed_samples 19700 unjoint_samples 19700 joint_samples 59 [1036772, 663728]
processed_samples 19701 unjoint_samples 19700 joint_samples 60 [185877, 1046295]
processed_samples 19700 unjoint_samples 19700 joint_samples 59 [930497, 1046332]
processed_samples 19700 unjoint_samples 19700 joint_samples 58 [872303, 1039287]
processed_samples 19701 unjoint_samples 19700 joint_samples 59 [1001129, 1046081]
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x559293470600] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a0e03700] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e79fb80] mmco: unref short failure
[h264 @ 0x55d97e79fb80] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97a69de40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a1f5b040] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x5592a4d279c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d96f79c540] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 19800 unjoint_samples 19800 joint_samples 60 [1047484, 145373]
[h264 @ 0x5592a168b680] mmco: unref short failure
processed_samples 19800 unjoint_samples 19800 joint_samples 59 [116216, 1039287]
processed_samples 19800 unjoint_samples 19800 joint_samples 59 [455144, 1047121]
processed_samples 19801 unjoint_samples 19800 joint_samples 60 [1044117, 309872]
processed_samples 19800 unjoint_samples 19800 joint_samples 58 [1047231, 740911]
processed_samples 19801 unjoint_samples 19800 joint_samples 60 [642117, 1046295]
processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1036772, 926301]
processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1047972, 694709]
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
processed_samples 19800 unjoint_samples 19800 joint_samples 60 [1047484, 145373]
processed_samples 19800 unjoint_samples 19800 joint_samples 59 [116216, 1039287]
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
processed_samples 19800 unjoint_samples 19800 joint_samples 59 [455144, 1047121]
processed_samples 19801 unjoint_samples 19800 joint_samples 60 [1044117, 309872]
processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1036772, 926301]
processed_samples 19800 unjoint_samples 19800 joint_samples 59 [1047972, 694709]
processed_samples 19801 unjoint_samples 19800 joint_samples 60 [642117, 1046295]
processed_samples 19800 unjoint_samples 19800 joint_samples 58 [1047231, 740911]
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x5592a1748140] mmco: unref short failure
[h264 @ 0x55d97e79fb80] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97eaf5040] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a534d800] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
processed_samples 19900 unjoint_samples 19900 joint_samples 60 [226377, 1026700]
processed_samples 19900 unjoint_samples 19900 joint_samples 59 [354550, 1039287]
processed_samples 19900 unjoint_samples 19900 joint_samples 59 [919999, 1047121]
processed_samples 19900 unjoint_samples 19900 joint_samples 58 [1047231, 999287]
processed_samples 19900 unjoint_samples 19900 joint_samples 60 [1047484, 424242]
processed_samples 19901 unjoint_samples 19900 joint_samples 60 [1041825, 1046295]
processed_samples 19901 unjoint_samples 19900 joint_samples 60 [1044117, 639618]
processed_samples 19900 unjoint_samples 19900 joint_samples 59 [1047972, 970929]
[h264 @ 0x5592a462d180] mmco: unref short failure
[h264 @ 0x5592a462d180] mmco: unref short failure
processed_samples 19900 unjoint_samples 19900 joint_samples 60 [226377, 1026700]
processed_samples 19900 unjoint_samples 19900 joint_samples 60 [1047484, 424242]
processed_samples 19900 unjoint_samples 19900 joint_samples 59 [919999, 1047121]
processed_samples 19900 unjoint_samples 19900 joint_samples 59 [354550, 1039287]
processed_samples 19900 unjoint_samples 19900 joint_samples 59 [1047972, 970929]
processed_samples 19901 unjoint_samples 19900 joint_samples 60 [1044117, 639618]
processed_samples 19901 unjoint_samples 19900 joint_samples 60 [1041825, 1046295]
processed_samples 19900 unjoint_samples 19900 joint_samples 58 [1047231, 999287]
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x55d97f86a7c0] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1f04380] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x5592a47dee40] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x55d97b9e9100] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x5592a49b9000] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x55d97e662e80] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97e08af80] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97c29fec0] mmco: unref short failure
[h264 @ 0x55d97c29fec0] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a50d5e80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d982fb5580] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
processed_samples 20000 unjoint_samples 20000 joint_samples 60 [288679, 1016785]
processed_samples 20000 unjoint_samples 20000 joint_samples 60 [609405, 1026700]
processed_samples 20000 unjoint_samples 20000 joint_samples 60 [288679, 1016785]
processed_samples 20000 unjoint_samples 20000 joint_samples 60 [1038668, 173023]
processed_samples 20000 unjoint_samples 20000 joint_samples 60 [1038668, 173023]
processed_samples 20000 unjoint_samples 20000 joint_samples 59 [233956, 1046742]
processed_samples 20000 unjoint_samples 20000 joint_samples 59 [233956, 1046742]
processed_samples 20000 unjoint_samples 20000 joint_samples 60 [1047484, 734847]
processed_samples 20000 unjoint_samples 20000 joint_samples 60 [609405, 1026700]
processed_samples 20000 unjoint_samples 20000 joint_samples 60 [1047484, 734847]
processed_samples 20001 unjoint_samples 20000 joint_samples 61 [21914, 1045160]
processed_samples 20001 unjoint_samples 20000 joint_samples 61 [21914, 1045160]
processed_samples 20001 unjoint_samples 20000 joint_samples 61 [1045681, 244261]
processed_samples 20001 unjoint_samples 20000 joint_samples 61 [1045681, 244261]
processed_samples 20000 unjoint_samples 20000 joint_samples 59 [784128, 1039287]
processed_samples 20000 unjoint_samples 20000 joint_samples 59 [784128, 1039287]
[h264 @ 0x55d97ac79900] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a099bac0] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x5592a4a4c140] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a0f1f600] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a45e4580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97c0d7140] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x55d97e5be6c0] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x5592a489d240] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
processed_samples 20100 unjoint_samples 20100 joint_samples 61 [115645, 1023199]
processed_samples 20100 unjoint_samples 20100 joint_samples 60 [962867, 1026700]
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
processed_samples 20100 unjoint_samples 20100 joint_samples 60 [1038668, 435728]
[h264 @ 0x5592a1125fc0] mmco: unref short failure
processed_samples 20100 unjoint_samples 20100 joint_samples 60 [49196, 1046542]
processed_samples 20100 unjoint_samples 20100 joint_samples 60 [651335, 1016785]
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
processed_samples 20101 unjoint_samples 20100 joint_samples 61 [333956, 1045160]
processed_samples 20100 unjoint_samples 20100 joint_samples 60 [49196, 1046542]
processed_samples 20101 unjoint_samples 20100 joint_samples 61 [1045681, 545026]
processed_samples 20100 unjoint_samples 20100 joint_samples 59 [587986, 1046742]
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
processed_samples 20100 unjoint_samples 20100 joint_samples 60 [1038668, 435728]
processed_samples 20100 unjoint_samples 20100 joint_samples 60 [651335, 1016785]
processed_samples 20100 unjoint_samples 20100 joint_samples 61 [115645, 1023199]
processed_samples 20100 unjoint_samples 20100 joint_samples 60 [962867, 1026700]
[h264 @ 0x55d97d7af000] mmco: unref short failure
processed_samples 20100 unjoint_samples 20100 joint_samples 59 [587986, 1046742]
processed_samples 20101 unjoint_samples 20100 joint_samples 61 [333956, 1045160]
processed_samples 20101 unjoint_samples 20100 joint_samples 61 [1045681, 545026]
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d982ab6fc0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d97b67efc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a4a42c80] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x5592a44f5fc0] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x55d97e573bc0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[mov,mp4,m4a,3gp,3g2,mj2 @ 0x55d979ccbf40] stream 1, offset 0x14007e3: partial file
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
processed_samples 20200 unjoint_samples 20200 joint_samples 61 [1029745, 293662]
processed_samples 20200 unjoint_samples 20200 joint_samples 61 [448020, 1023199]
processed_samples 20200 unjoint_samples 20200 joint_samples 59 [896201, 1046742]
processed_samples 20200 unjoint_samples 20200 joint_samples 60 [349592, 1046542]
processed_samples 20201 unjoint_samples 20200 joint_samples 60 [1038668, 725572]
processed_samples 20200 unjoint_samples 20200 joint_samples 60 [1009376, 1016785]
processed_samples 20201 unjoint_samples 20200 joint_samples 61 [661225, 1045160]
processed_samples 20201 unjoint_samples 20200 joint_samples 61 [1045681, 881604]
[h264 @ 0x5592a45e4580] mmco: unref short failure
[mov,mp4,m4a,3gp,3g2,mj2 @ 0x5592a0a1f2c0] stream 1, offset 0x14007e3: partial file
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x5592a5128b40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
processed_samples 20200 unjoint_samples 20200 joint_samples 61 [1029745, 293662]
processed_samples 20200 unjoint_samples 20200 joint_samples 61 [448020, 1023199]
processed_samples 20200 unjoint_samples 20200 joint_samples 60 [349592, 1046542]
processed_samples 20200 unjoint_samples 20200 joint_samples 60 [1009376, 1016785]
processed_samples 20201 unjoint_samples 20200 joint_samples 61 [661225, 1045160]
processed_samples 20200 unjoint_samples 20200 joint_samples 59 [896201, 1046742]
processed_samples 20201 unjoint_samples 20200 joint_samples 60 [1038668, 725572]
processed_samples 20201 unjoint_samples 20200 joint_samples 61 [1045681, 881604]
[h264 @ 0x5592a2973f40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2973f40] mmco: unref short failure
[h264 @ 0x5592a2973f40] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a2973f40] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x55d97bab4b80] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97c863300] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a0b8bd40] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97e323800] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a118ff40] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x5592a1cff4c0] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97a495540] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d97be9d240] mmco: unref short failure
[h264 @ 0x55d9818fc2c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x55d97d921300] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x5592a4e9f6c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a39e8940] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a51031c0] mmco: unref short failure
[h264 @ 0x5592a4fb2540] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
processed_samples 20300 unjoint_samples 20300 joint_samples 61 [1029745, 533251]
[h264 @ 0x5592a529c8c0] mmco: unref short failure
processed_samples 20300 unjoint_samples 20300 joint_samples 60 [1038179, 206890]
processed_samples 20300 unjoint_samples 20300 joint_samples 61 [1025310, 266540]
processed_samples 20300 unjoint_samples 20300 joint_samples 61 [802462, 1023199]
processed_samples 20300 unjoint_samples 20300 joint_samples 60 [654416, 1046542]
processed_samples 20301 unjoint_samples 20300 joint_samples 61 [881264, 1045160]
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
processed_samples 20301 unjoint_samples 20300 joint_samples 62 [279304, 988659]
[h264 @ 0x55d97be9d240] mmco: unref short failure
processed_samples 20301 unjoint_samples 20300 joint_samples 60 [1038668, 1022620]
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x55d97aa0c640] mmco: unref short failure
processed_samples 20300 unjoint_samples 20300 joint_samples 61 [1029745, 533251]
processed_samples 20300 unjoint_samples 20300 joint_samples 60 [1038179, 206890]
processed_samples 20300 unjoint_samples 20300 joint_samples 61 [1025310, 266540]
processed_samples 20300 unjoint_samples 20300 joint_samples 61 [802462, 1023199]
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
processed_samples 20300 unjoint_samples 20300 joint_samples 60 [654416, 1046542]
processed_samples 20301 unjoint_samples 20300 joint_samples 62 [279304, 988659]
processed_samples 20301 unjoint_samples 20300 joint_samples 60 [1038668, 1022620]
processed_samples 20301 unjoint_samples 20300 joint_samples 61 [881264, 1045160]
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x5592a215f4c0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x55d981fe2cc0] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a2270c00] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x55d97b490480] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x55d97b5edb00] mmco: unref short failure
[h264 @ 0x55d97b5edb00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x55d97adf2580] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97e1e1900] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a1e31540] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a10ae380] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x55d97b701480] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
[h264 @ 0x5592a45d9d80] mmco: unref short failure
processed_samples 20400 unjoint_samples 20400 joint_samples 62 [148440, 1025591]
processed_samples 20400 unjoint_samples 20400 joint_samples 60 [1038179, 499918]
processed_samples 20401 unjoint_samples 20400 joint_samples 62 [175409, 1047339]
processed_samples 20401 unjoint_samples 20400 joint_samples 61 [1046467, 263572]
processed_samples 20401 unjoint_samples 20400 joint_samples 62 [595708, 988659]
processed_samples 20400 unjoint_samples 20400 joint_samples 61 [1025310, 610443]
processed_samples 20400 unjoint_samples 20400 joint_samples 61 [1029745, 840603]
processed_samples 20400 unjoint_samples 20400 joint_samples 60 [922462, 1046542]
[h264 @ 0x55d97b701480] mmco: unref short failure
processed_samples 20400 unjoint_samples 20400 joint_samples 62 [148440, 1025591]
[h264 @ 0x55d97b930d80] mmco: unref short failure
[h264 @ 0x55d97b930d80] mmco: unref short failure
processed_samples 20400 unjoint_samples 20400 joint_samples 61 [1025310, 610443]
processed_samples 20400 unjoint_samples 20400 joint_samples 60 [1038179, 499918]
processed_samples 20401 unjoint_samples 20400 joint_samples 62 [175409, 1047339]
processed_samples 20400 unjoint_samples 20400 joint_samples 61 [1029745, 840603]
processed_samples 20401 unjoint_samples 20400 joint_samples 61 [1046467, 263572]
processed_samples 20401 unjoint_samples 20400 joint_samples 62 [595708, 988659]
[h264 @ 0x55d97a5b2680] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
processed_samples 20400 unjoint_samples 20400 joint_samples 60 [922462, 1046542]
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a0bedc40] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x55d97f274580] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97e3ea240] mmco: unref short failure
[h264 @ 0x55d97ae8c5c0] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97ea3b980] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a0dc4f00] mmco: unref short failure
[h264 @ 0x5592a4da9c00] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a529c8c0] mmco: unref short failure
[h264 @ 0x5592a0992b40] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x55d97a4b9940] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
processed_samples 20500 unjoint_samples 20500 joint_samples 62 [1043854, 97683]
processed_samples 20500 unjoint_samples 20500 joint_samples 62 [1043854, 97683]
processed_samples 20500 unjoint_samples 20500 joint_samples 62 [464899, 1025591]
processed_samples 20500 unjoint_samples 20500 joint_samples 61 [1025310, 858603]
processed_samples 20500 unjoint_samples 20500 joint_samples 61 [1032513, 368428]
processed_samples 20500 unjoint_samples 20500 joint_samples 61 [1032513, 368428]
processed_samples 20501 unjoint_samples 20500 joint_samples 62 [508923, 1047339]
processed_samples 20501 unjoint_samples 20500 joint_samples 61 [1046467, 734134]
processed_samples 20500 unjoint_samples 20500 joint_samples 62 [464899, 1025591]
processed_samples 20501 unjoint_samples 20500 joint_samples 61 [1046467, 734134]
processed_samples 20501 unjoint_samples 20500 joint_samples 62 [916831, 988659]
processed_samples 20501 unjoint_samples 20500 joint_samples 62 [916831, 988659]
processed_samples 20501 unjoint_samples 20500 joint_samples 62 [508923, 1047339]
processed_samples 20500 unjoint_samples 20500 joint_samples 60 [1038179, 930073]
processed_samples 20500 unjoint_samples 20500 joint_samples 60 [1038179, 930073]
processed_samples 20500 unjoint_samples 20500 joint_samples 61 [1025310, 858603]
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x5592a513f580] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x55d97ac67100] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x55d9826bc180] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x5592a4e8d900] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x5592a31fac00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a10b57c0] mmco: unref short failure
[h264 @ 0x5592a0e40300] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x55d97b4bbf80] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x5592a2fab840] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x55d97b455300] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x5592a23df380] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x55d97b543c80] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a1060b80] mmco: unref short failure
[h264 @ 0x5592a0bf6640] mmco: unref short failure
[h264 @ 0x55d9819618c0] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97e64a980] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97b679340] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
processed_samples 20600 unjoint_samples 20600 joint_samples 62 [52850, 1046863]
processed_samples 20600 unjoint_samples 20600 joint_samples 62 [1043854, 470938]
processed_samples 20600 unjoint_samples 20600 joint_samples 61 [1044262, 278088]
processed_samples 20600 unjoint_samples 20600 joint_samples 61 [1032513, 661476]
processed_samples 20600 unjoint_samples 20600 joint_samples 62 [796788, 1025591]
[h264 @ 0x55d97e1f6c00] mmco: unref short failure
[h264 @ 0x5592a0447f80] mmco: unref short failure
[h264 @ 0x5592a0447f80] mmco: unref short failure
processed_samples 20601 unjoint_samples 20600 joint_samples 63 [1038226, 211867]
processed_samples 20601 unjoint_samples 20600 joint_samples 62 [801183, 1047339]
processed_samples 20600 unjoint_samples 20600 joint_samples 62 [52850, 1046863]
[h264 @ 0x55d97eb10a00] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
[h264 @ 0x5592a48560c0] mmco: unref short failure
processed_samples 20600 unjoint_samples 20600 joint_samples 62 [1043854, 470938]
processed_samples 20600 unjoint_samples 20600 joint_samples 61 [1032513, 661476]
processed_samples 20601 unjoint_samples 20600 joint_samples 61 [1046467, 1025343]
processed_samples 20601 unjoint_samples 20600 joint_samples 61 [1046467, 1025343]
processed_samples 20600 unjoint_samples 20600 joint_samples 61 [1044262, 278088]
[h264 @ 0x5592a0447f80] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
processed_samples 20601 unjoint_samples 20600 joint_samples 63 [1038226, 211867]
processed_samples 20600 unjoint_samples 20600 joint_samples 62 [796788, 1025591]
processed_samples 20601 unjoint_samples 20600 joint_samples 62 [801183, 1047339]
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a18a8cc0] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97acc3440] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x55d97c0e2ac0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a1125fc0] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a12fb640] mmco: unref short failure
[h264 @ 0x5592a522b380] mmco: unref short failure
[h264 @ 0x55d97a978c40] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a2142e80] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x5592a1fbe040] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97c8919c0] mmco: unref short failure
[h264 @ 0x55d97d7af000] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97a987100] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x55d97b3130c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a17c95c0] mmco: unref short failure
[h264 @ 0x5592a4deb500] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a70bec80] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a21a1900] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a487d5c0] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x55d98309b480] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a168b680] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x5592a2a46600] mmco: unref short failure
[h264 @ 0x55d97f7f4a40] mmco: unref short failure
[h264 @ 0x55d97a5c0900] mmco: unref short failure
[h264 @ 0x5592a0e44380] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e32fd00] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a1a6da40] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d979dbe500] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
[h264 @ 0x55d97c3b4980] mmco: unref short failure
processed_samples 20700 unjoint_samples 20700 joint_samples 62 [311420, 1046863]
processed_samples 20700 unjoint_samples 20700 joint_samples 61 [1044262, 630548]
processed_samples 20700 unjoint_samples 20700 joint_samples 61 [1032513, 1019131]
processed_samples 20700 unjoint_samples 20700 joint_samples 62 [1030443, 1031633]
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
[h264 @ 0x55d97b4be540] mmco: unref short failure
processed_samples 20701 unjoint_samples 20700 joint_samples 62 [1046467, 281359]
processed_samples 20700 unjoint_samples 20700 joint_samples 62 [1043854, 766145]
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d97b4ca980] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
processed_samples 20701 unjoint_samples 20700 joint_samples 63 [1038226, 479425]
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
processed_samples 20701 unjoint_samples 20700 joint_samples 63 [1046432, 109341]
[h264 @ 0x55d97a987100] mmco: unref short failure
processed_samples 20700 unjoint_samples 20700 joint_samples 62 [311420, 1046863]
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x5592a183a9c0] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
[h264 @ 0x55d97b84ba80] mmco: unref short failure
processed_samples 20701 unjoint_samples 20700 joint_samples 62 [1046467, 281359]
processed_samples 20700 unjoint_samples 20700 joint_samples 62 [1043854, 766145]
processed_samples 20700 unjoint_samples 20700 joint_samples 62 [1030443, 1031633]
processed_samples 20700 unjoint_samples 20700 joint_samples 61 [1044262, 630548]
processed_samples 20700 unjoint_samples 20700 joint_samples 61 [1032513, 1019131]
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
processed_samples 20701 unjoint_samples 20700 joint_samples 63 [1038226, 479425]
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
processed_samples 20701 unjoint_samples 20700 joint_samples 63 [1046432, 109341]
[h264 @ 0x5592a4a64840] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a2305440] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x5592a55fb000] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a9a59e00] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a4a39d40] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x55d97e0120c0] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x5592a1035240] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d981aabf40] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97b321300] mmco: unref short failure
[h264 @ 0x55d97ba0ddc0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a48b0ac0] mmco: unref short failure
[h264 @ 0x5592a2a54580] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x5592a1a01d80] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a16e1b00] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a0d96180] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x55d97ee241c0] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a2682000] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x55d97c6b99c0] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a1f2c740] mmco: unref short failure
[h264 @ 0x5592a13f58c0] mmco: unref short failure
[h264 @ 0x5592a0bce4c0] mmco: unref short failure
[h264 @ 0x55d97aa12080] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x5592a1c74200] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d979c2f500] mmco: unref short failure
[h264 @ 0x55d97e6e3d80] mmco: unref short failure
[h264 @ 0x5592a22c0d80] mmco: unref short failure
processed_samples 20800 unjoint_samples 20800 joint_samples 63 [1048090, 234977]
processed_samples 20800 unjoint_samples 20800 joint_samples 63 [125204, 1008335]
processed_samples 20800 unjoint_samples 20800 joint_samples 63 [1048090, 234977]
processed_samples 20800 unjoint_samples 20800 joint_samples 63 [125204, 1008335]
processed_samples 20800 unjoint_samples 20800 joint_samples 62 [574724, 1046863]
processed_samples 20800 unjoint_samples 20800 joint_samples 62 [248990, 1037734]
processed_samples 20800 unjoint_samples 20800 joint_samples 62 [574724, 1046863]
processed_samples 20800 unjoint_samples 20800 joint_samples 62 [248990, 1037734]
processed_samples 20801 unjoint_samples 20800 joint_samples 62 [1046467, 589561]
processed_samples 20801 unjoint_samples 20800 joint_samples 62 [1046467, 589561]
processed_samples 20800 unjoint_samples 20800 joint_samples 61 [1044262, 876985]
processed_samples 20801 unjoint_samples 20800 joint_samples 63 [1046432, 398225]
processed_samples 20801 unjoint_samples 20800 joint_samples 63 [1038226, 978763]
processed_samples 20801 unjoint_samples 20800 joint_samples 63 [1046432, 398225]
[h264 @ 0x5592a08edc00] mmco: unref short failure
[h264 @ 0x5592a08edc00] mmco: unref short failure
processed_samples 20800 unjoint_samples 20800 joint_samples 61 [1044262, 876985]
processed_samples 20801 unjoint_samples 20800 joint_samples 63 [1038226, 978763]
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x5592a118fa40] mmco: unref short failure
[h264 @ 0x55d97b459a00] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x55d97dd14f40] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4e100c0] mmco: unref short failure
[h264 @ 0x5592a4810c40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97e049e00] mmco: unref short failure
[h264 @ 0x55d97af04e00] mmco: unref short failure
[h264 @ 0x55d98337cc80] mmco: unref short failure
[h264 @ 0x5592a0a98580] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x55d982fa7a40] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[h264 @ 0x5592a19044c0] mmco: unref short failure
[2024-12-01 22:44:43,471] torch.distributed.elastic.agent.server.api: [ERROR] Error waiting on exit barrier. Elapsed: 300.10550570487976 seconds
+ set +x