SmolVLM-256M is the smallest multimodal model in the world. It accepts arbitrary sequences of image and text inputs to produce text outputs. It's designed for efficiency. SmolVLM can answer questions about images, describe visual content, or transcribe text. Its lightweight architecture makes it suitable for on-device applications while maintaining strong performance on multimodal tasks. It can run inference on one image with under 1GB of GPU RAM.
The Base Model providing a 128 context window and a maximum output of 1,024 tokens.
Support Platforms: LLM630 Compute Kit, Module LLM, and Module LLM Kit
apt install llm-model-smolvlm-256m-ax630c