Llama 4 Scout (free) Check detailed information and pricing for AI models

Context Length 128,000 tokens, meta-llama from provided

128,000
Context Tokens
Free
Prompt Price
Free
Output Price
6/16
Feature Support

Model Overview

Llama 4 Scout 17B Instruct (16E) is a mixture-of-experts (MoE) language model developed by Meta, activating 17 billion parameters out of a total of 109B. It supports native multimodal input (text and image) and multilingual output (text and code) across 12 supported languages. Designed for assistant-style interaction and visual reasoning, Scout uses 16 experts per forward pass and features a context length of 10 million tokens, with a training corpus of ~40 trillion tokens. Built for high efficiency and local or commercial deployment, Llama 4 Scout incorporates early fusion for seamless modality integration. It is instruction-tuned for use in multilingual chat, captioning, and image understanding tasks. Released under the Llama 4 Community License, it was last trained on data up to August 2024 and launched publicly on April 5, 2025.

Basic Information

Developer
meta-llama
Model Series
Llama4
Release Date
2025-04-05
Context Length
128,000 tokens
Max Completion Tokens
4,028 tokens
Variant
free

Pricing Information

This model is free to use

Supported Features

Supported (6)

Image Input
Top K
Repetition Penalty
Response Format
Tool Usage
Structured Outputs

Unsupported (10)

Seed
Frequency Penalty
Presence Penalty
Min P
Logit Bias
Logprobs
Top Logprobs
Reasoning
Web Search Options
Top A

Other Variants

Actual Usage Statistics

No recent usage data available.

Models by Same Author (meta-llama)

Llama 3.3 8B Instruct (free)
128,000 tokens
Free
Llama 3.3 8B Instruct
128,000 tokens
$0.00 / $0.00
Llama Guard 4 12B
163,840 tokens
$0.18 / $0.18
Llama 4 Maverick (free)
128,000 tokens
Free
Llama 4 Maverick
1,048,576 tokens
$0.15 / $0.60