Meta: Llama 3.2 11B Vision Instruct
Meta: Llama 3.2 11B Vision Instruct is a multimodal model for vision-language understanding. It combines multimodal input handling and image understanding with a 131K tokens context window and a low-cost profile. Use it for vision-language understanding and content analysis when quality, speed, and cost matters.
Input
$0.24/1M
Output
$0.24/1M
Cached
$0.01/1M
Batch
$0.03/1M