Meta-Llama-3.1-8B-Instruct

Meta-Llama-3.1-8B-Instruct

meta-llama/Meta-Llama-3.1-8B-Instruct

About Meta-Llama-3.1-8B-Instruct

Meta Llama 3.1 is a family of multilingual large language models developed by Meta, featuring pretrained and instruction-tuned variants in 8B, 70B, and 405B parameter sizes. This 8B instruction-tuned model is optimized for multilingual dialogue use cases and outperforms many available open-source and closed chat models on common industry benchmarks. The model was trained on over 15 trillion tokens of publicly available data, using techniques like supervised fine-tuning and reinforcement learning with human feedback to enhance helpfulness and safety. Llama 3.1 supports text and code generation, with a knowledge cutoff of December 2023

Available Serverless

Run queries immediately, pay only for usage

$

0.06

/

$

0.06

Per 1M Tokens (input/output)

Metadata

Create on

Apr 23, 2025

License

Provider

Meta Llama

Specification

State

Available

Architecture

Calibrated

Yes

Mixture of Experts

No

Total Parameters

8

Activated Parameters

8B

Reasoning

No

Precision

FP8

Context length

33K

Max Tokens

4K

Supported Functionality

Serverless

Supported

Serverless LoRA

Not supported

Fine-tuning

Not supported

Embeddings

Not supported

Rerankers

Not supported

Support image input

Not supported

JSON Mode

Supported

Structured Outputs

Not supported

Tools

Not supported

Fim Completion

Not supported

Chat Prefix Completion

Supported

Model FAQs: Usage, Deployment

Learn how to use, fine-tune, and deploy this model with ease.

Ready to accelerate your AI development?

Ready to accelerate your AI development?