Context
131K tokens (~66 books)
Input $/1M
$0.24
Output $/1M
$0.24
Type
multimodal
License
Open Source
Benchmarks
0 tested
Data updated today
About
Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and...
No benchmark data available yet.
Links
Research
Documentation
Community
Source Code
BenchGecko API
llama-3-2-11b-vision-instruct
Specifications
- Typemultimodal
- Context131K tokens (~66 books)
- ReleasedSep 2024
- LicenseOpen Source
- StatusActive
- Cost / Message~$0.001
Available On
Share & Export
Frequently Asked Questions
Llama 3.2 11B Vision Instruct is an open-source multimodal AI model by Meta, released in September 2024. Context window: 131K tokens.