Skip to content
Azure AI

Llama 3.2 90B Vision Instruct

Llama 3.2 90B Vision Instruct is available via Azure AI with a 128K context window and up to 2,048 output tokens. Pricing: $2.04/1M input tokens, $2.04/1M output tokens.

Llama 3.2 90B Vision Instruct Pricing & Specifications

Input Price$2.04 per 1M tokens
Output Price$2.04 per 1M tokens
Context Window128,000 tokens (128K)
Max Output2,048 tokens
ProviderAzure AI

What is Llama 3.2 90B Vision Instruct?

Llama 3.2 90B Vision Instruct is a large language model by Azure AI with a 128K context window and up to 2,048 output tokens. It costs $2.04 per 1M input tokens and $2.04 per 1M output tokens. Llama 3.2 90B Vision Instruct is available via Azure AI with a 128K context window and up to 2,048 output tokens. Pricing: $2.04/1M input tokens, $2.04/1M output tokens.

Capabilities

text vision function calling

Llama 3.2 90B Vision Instruct Cost Examples

Short prompt (500 tokens)

$0.001020

Medium prompt (2K tokens)

$0.00408

Long output (4K tokens)

$0.00816

Count tokens for Llama 3.2 90B Vision Instruct

Paste your prompt to see exact token counts and API cost estimates.

Open Token Counter

Similar Models to Llama 3.2 90B Vision Instruct

Azure AI

Mistral Large 2407

$2.00/1M in 128K ctx

Azure AI

Mistral Large Latest

$2.00/1M in 128K ctx

Azure AI

Llama 4 Maverick 17B 128E Instruct FP8

$1.41/1M in 1M ctx

Azure AI

Meta Llama 3.1 70B Instruct

$2.68/1M in 128K ctx

Frequently Asked Questions

How much does Llama 3.2 90B Vision Instruct cost per token? +
Llama 3.2 90B Vision Instruct costs $2.04 per 1M input tokens and $2.04 per 1M output tokens. For a typical 1,000-token request with a 500-token response, that works out to roughly $0.003060.
What is the context window for Llama 3.2 90B Vision Instruct? +
Llama 3.2 90B Vision Instruct supports a context window of 128,000 tokens (128K). This determines the maximum combined length of your prompt and conversation history in a single API call.
What is the maximum output length for Llama 3.2 90B Vision Instruct? +
Llama 3.2 90B Vision Instruct can generate up to 2,048 tokens in a single response. If you need longer outputs, you can make multiple API calls and concatenate the results.
Is Llama 3.2 90B Vision Instruct good for coding tasks? +
Yes, Llama 3.2 90B Vision Instruct supports capabilities well-suited for coding tasks including code generation, debugging, and refactoring.
Token Counter | Pricing Calculator | Model Comparison | All Azure AI Models