DOSDOS
Pricing
Get started
Models/Llama 3.3 70B

Llama 3.3 70B

New

by Meta

High-performance multilingual LLM optimized for dialogue and instruction following.

Parameters
70B
Context Length
128K
Category
chat
Coming Soon

This model is not yet available. Stay tuned for updates.

Browse available models

About this model

Llama 3.3 70B is Meta's latest instruction-tuned language model, offering exceptional performance across a wide range of tasks. With 70 billion parameters and a 128K context window, it excels at complex reasoning, coding, multilingual tasks, and creative writing. The model has been fine-tuned using RLHF to be helpful, harmless, and honest.

Capabilities

Chat & DialogueInstruction FollowingCode GenerationMultilingualReasoningFunction Calling

Use Cases

  • Customer Support Bots
  • Code Assistants
  • Content Generation
  • Data Analysis
  • Research Assistants

Model Details

Provider
Meta
Model ID
meta-llama/Llama-3.3-70B-Instruct
Parameters
70B
Context Length
128K tokens
Category
chat

API Usage

Use the DOS API to integrate Llama 3.3 70Binto your applications. Our API is compatible with OpenAI's client libraries for easy migration.

Model ID

meta-llama/Llama-3.3-70B-Instruct

Python

python
from dos import DOS

client = DOS()

response = client.chat.completions.create(
    model="meta-llama/Llama-3.3-70B-Instruct",
    messages=[
        {"role": "user", "content": "Hello, how are you?"}
    ]
)

print(response.choices[0].message.content)

cURL

bash
curl https://api.dos.ai/v1/chat/completions \
  -H "Authorization: Bearer $DOS_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "meta-llama/Llama-3.3-70B-Instruct",
    "messages": [
      {"role": "user", "content": "Hello, how are you?"}
    ]
  }'

Node.js

javascript
import DOS from 'dos-ai';

const client = new DOS();

const response = await client.chat.completions.create({
  model: "meta-llama/Llama-3.3-70B-Instruct",
  messages: [
    { role: "user", content: "Hello, how are you?" }
  ]
});

console.log(response.choices[0].message.content);
View full API reference

Related Models

Qwen

Qwen 3.5 35B-A3B

Ultra-efficient MoE model — 35B total, 3B active parameters. Fast inference at near-8B cost with 70B-class quality.

$0.50 / 1M tokens
Meta

Llama 3.1 405B

The largest and most capable Llama model for complex reasoning and generation tasks.

$3.50 / 1M tokens
Mistral AI

Mistral Large 2

Flagship model with strong multilingual and coding capabilities.

$2.00 / 1M tokens
DOSDOS

AI infrastructure for everyone. Inference, agents, and safety — all in one platform.

Product

  • Models
  • Pricing
  • API Inference
  • DOSClaw
  • GPU Cloud

Developers

  • Documentation
  • API Reference
  • Status

DOS Ecosystem

  • DOSafe
  • DOS.Me
  • DOScan
  • DOSwap
  • MetaDOS

Company

  • About
  • Contact
  • Careers
  • Privacy
  • Terms

© 2026 All rights reserved.