Home/Models/Llama 3.2 11B Vision Instruct
Meta logo

Llama 3.2 11B Vision Instruct

by Meta · Released Sep 2024

Open SourceMultimodal
Compare
Context
131K tokens (~66 books)
Input $/1M
$0.24
Output $/1M
$0.24
Type
multimodal
License
Open Source
Benchmarks
0 tested
Data updated today
About

Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and...

No benchmark data available yet.

Links
Documentation
Community
BenchGecko API
llama-3-2-11b-vision-instruct
Specifications
  • Typemultimodal
  • Context131K tokens (~66 books)
  • ReleasedSep 2024
  • LicenseOpen Source
  • StatusActive
  • Cost / Message~$0.001
Available On
Meta logoMeta$0.24
Share & Export
Tweet
Llama 3.2 11B Vision Instruct is an open-source multimodal AI model by Meta, released in September 2024. Context window: 131K tokens.