← Back to model catalog

Llama 3.3 70B Instruct

Provider
Brightnode-hosted
Context
131,072 tokens
Pricing
$0.22 / 1M input · $0.50 / 1M output
APAC regions
Singapore
Residency
in-region
Task
Chat
APAC performance

Latency profile

SingaporeTokyoSydney
TTFT p50274134
TTFT p95446657
E2E latency p50627871
E2E latency p9596121109
Pricing

Input: $0.22 per 1M tokens

Output: $0.50 per 1M tokens

Billing: Per-token, charged against wallet balance

Dedicated endpoint option
  • A100 80GB: $4.01/hr (Singapore)
  • H100 80GB: $14.29/hr (Singapore)
View full pricing and calculator →
Capabilities

Meta's multilingual instruction model with strong chat and reasoning quality for long-context workflows.

Best for: General chat, Code assistance, Multilingual APAC use cases, RAG pipelines, Agent backends

Quickstart code snippets
Python
from openai import OpenAI

client = OpenAI(
    base_url="https://api.brightnode.cloud/v1",
    api_key="YOUR_BRIGHTNODE_API_KEY",
)

response = client.chat.completions.create(
    model="meta-llama/Llama-3.3-70B-Instruct",
    messages=[{"role": "user", "content": "Hello from APAC"}],
)
Node
import OpenAI from "openai";

const client = new OpenAI({
  baseURL: "https://api.brightnode.cloud/v1",
  apiKey: process.env.BRIGHTNODE_API_KEY,
});

const response = await client.chat.completions.create({
  model: "meta-llama/Llama-3.3-70B-Instruct",
  messages: [{ role: "user", content: "Hello from APAC" }],
});
Other models on Brightnode