Deploy your private API instance of Llama3.1 via the VLLM inference library.
Select one of the pre-configured Llama VLLM
Deploy your GPU instance and wait until the installation process is complete
Click our Port Information Button
Follow the Instructions to connect to your Jupyter Notebook
Deploy your private API instance of Llama3.1 via the VLLM inference library.
Select one of the pre-configured Llama VLLM
Deploy your GPU instance and wait until the installation process is complete
Click our Port Information Button
Follow the Instructions to connect to your Jupyter Notebook