Cutting-edge AI inference and training, unmatched cloud-native experience, and top-tier GPU infrastructure.
Efficient, reliable and easy to use
Combining high performance computing with cloud native efficiency
We built the fastest and scalable AI runtimes
# Installpip install -U leptonai# Serve huggingface modellep photon run -n llama3 -m hf:meta-llama/Meta-Llama-3-8B-Instruct# Serve vllm modellep photon run -n mixtral -m vllm:mistralai/Mixtral-8x7B-v0.1# Serve with Tuna, Lepton's optimized engine (coming soon!)lep tuna run -n mixtral -m mistralai/Mistral-7B-Instruct-v0.3
High performance computation hardware and cloud native software combined