# 🌟 New adapter setup ## Model description Llama-3.2-11B-Vision-Instruct is multimodal LLM that allows to have images and text as input. ## Open source status * [x] the model implementation is available: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct * [x] the model weights are available: https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct * [x] Authors: Meta