When I inference video based internvl model, I find the frames of video is not transformer in vllm and only there are 32 frames. Therefore, I want to know how to set num_frames for video. I try this server: --media-io-kwargs '{"video": {"num_frames": 60}}' but I don't konw if it is right.
env CUDA_VISIBLE_DEVICES=2 vllm serve /mnt/models/InternVL3_5-4B \
-dp 1 \
--gpu-memory-utilization 0.8 \
--max-model-len 40K \
--served-model-name InternVL3_5-4B \
--media-io-kwargs '{"video": {"num_frames": 60}}' \
--port 12400 \
--trust-remote-code
``