Mixtral 8x7b
32K context
Description
The Mixtral 8x7b Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts.
Pricing
- Dedicated Endpoints: Calculated by the instance type and the number of GPUs, you can find the details in pricing page. You can also contact us to reserve GPUs.
- Serverless Endpoints: $0.5 / M tokens for using Mixtral 8x7b, pay as you go.
Playground
System prompt
Temperature
Max tokens
Top P