meta-llama

Meta: Llama 3.3 70B Instruct (free)

Meta: Llama 3.3 70B Instruct (free) is the free-tier variant of meta-llama's Llama 3.3 70B Instruct model, available at $0 per million tokens for both input and output. It is a text-to-text model with a 65,536-token context window. This specific variant has not been included in our standard benchmark testing suite; however, the paid Llama 3.3 70B Instruct has been tested and scored an average of 3.5 across our 12-benchmark suite — ranking 43rd out of 52 models overall. The free variant provides zero-cost access to the same underlying model, typically subject to rate limits. Within the meta-llama family, the only other tested sibling is Llama 4 Scout (avg 3.33, $0.30/M output), which the paid Llama 3.3 70B Instruct outperforms on average.

Performance

We have not benchmarked Meta: Llama 3.3 70B Instruct (free) directly in our 12-test suite. The paid Llama 3.3 70B Instruct has been tested and scored an average of 3.5 across our benchmarks, placing it 43rd out of 52 models overall. For detailed per-benchmark scores and rankings, refer to the Llama 3.3 70B Instruct profile. The free variant is expected to perform equivalently at the model level, with rate limits as the primary practical difference. Our tested set median is above 3.5, so the paid variant places in the lower half of the field on average, though individual benchmark results vary.

Pricing

This is a free model — $0 per million input and output tokens. Free-tier access typically carries rate limits and lower queuing priority compared to paid access. The paid Llama 3.3 70B Instruct costs $0.32/M output — at 10 million output tokens per month, that's $3.20; at 100 million tokens, $32.00. For developers and teams evaluating the model before a production commitment, the free tier provides a zero-cost path. For high-throughput production use, the paid variant ensures more reliable availability. At $0.32/M output, the paid Llama 3.3 70B Instruct is priced competitively against other mid-tier models in the tested set.

meta-llama

Meta: Llama 3.3 70B Instruct (free)

Overall
0.00/5N/A

External Benchmarks

SWE-bench Verified
N/A
MATH Level 5
N/A
AIME 2025
N/A

Pricing

Input

$0.00/MTok

Output

$0.00/MTok

Context Window66K

modelpicker.net

Real-World Costs

iChat response$0.00
iBlog post$0.00
iDocument batch$0.00
iPipeline run$0.00

Pricing vs Performance

Output cost per million tokens (log scale) vs average score across our 12 internal benchmarks

This modelOther models

Try It

from openai import OpenAI

client = OpenAI(
    base_url="https://openrouter.ai/api/v1",
    api_key="YOUR_OPENROUTER_KEY",
)

response = client.chat.completions.create(
    model="meta-llama/llama-3.3-70b-instruct:free",
    messages=[
        {"role": "user", "content": "Hello, Meta: Llama 3.3 70B Instruct (free)!"}
    ],
)

print(response.choices[0].message.content)

Recommendation

Meta: Llama 3.3 70B Instruct (free) is a good starting point for developers who want zero-cost access to a 70B-parameter text model for prototyping, experimentation, and low-volume use. The free tier allows full evaluation of the model's capabilities before deciding whether to move to the paid variant at $0.32/M output. With the paid Llama 3.3 70B Instruct scoring 3.5 on average across our benchmarks (rank 43 of 52), it is a mid-tier model — capable for general tasks but not a top performer in reasoning-intensive or safety-calibrated use cases. For teams that need a free text model at zero cost and are comfortable with rate limits, this is a viable option. Teams needing higher overall benchmark performance within a comparable cost tier should compare the paid Llama 3.3 70B Instruct ($0.32/M output) against other options in the $0.30-$0.40 output range.

How We Test

We test every model against our 12-benchmark suite covering tool calling, agentic planning, creative problem solving, safety calibration, and more. Each test is scored 1–5 by an LLM judge. Read our full methodology.