On-Demand Cloud
Deploy Llama via VLLM
Deploy your private API instance of Llama3.1 via the VLLM inference library.
Step-by-Step Guide
1
Select one of the pre-configured Llama VLLM
2
Deploy your GPU instance and wait until the installation process is complete
3
Click our Port Information Button
4
Follow the Instructions to connect to your Jupyter Notebook