Mixtral 8x7b

32K context

Description

The Mixtral 8x7b Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts.

Pricing

  • Dedicated Endpoints: Calculated by the instance type and the number of GPUs, you can find the details in pricing page. You can also contact us to reserve GPUs.
  • Serverless Endpoints: $0.5 / M tokens for using Mixtral 8x7b, pay as you go.

Playground

System prompt
Temperature
Max tokens
Top P

API Reference

Lepton AI

© 2025