POST
/
api
/
v1
/
pods
curl --request POST \
  --url https://api.primeintellect.ai/api/v1/pods/ \
  --header 'Authorization: Bearer <token>' \
  --header 'Content-Type: application/json' \
  --data '{
  "pod": {
    "name": "<string>",
    "cloudId": "<string>",
    "gpuType": "CPU_NODE",
    "socket": "PCIe",
    "gpuCount": 123,
    "diskSize": 123,
    "vcpus": 123,
    "memory": 123,
    "maxPrice": 123,
    "image": "ubuntu_22_cuda_12",
    "customTemplateId": "<string>",
    "dataCenterId": "<string>",
    "country": "<string>",
    "security": "secure_cloud",
    "jupyterPassword": "<string>",
    "autoRestart": true
  },
  "provider": {
    "type": "runpod"
  },
  "team": {
    "teamId": "<string>"
  }
}'
{
  "id": "<string>",
  "userId": "<string>",
  "teamId": "<string>",
  "walletId": "<string>",
  "name": "<string>",
  "type": "HOSTED",
  "providerType": "runpod",
  "status": "PROVISIONING",
  "installationStatus": "PENDING",
  "installationFailure": "<string>",
  "installationProgress": 123,
  "createdAt": "2023-11-07T05:31:56Z",
  "updatedAt": "2023-11-07T05:31:56Z",
  "gpuName": "CPU_NODE",
  "gpuCount": 1,
  "jupyterPassword": "<string>",
  "socket": "PCIe",
  "priceHr": 1.23,
  "stoppedPriceHr": 0.005,
  "provisioningPriceHr": 123,
  "basePriceHr": 123,
  "baseCurrency": "<string>",
  "environmentType": "ubuntu_22_cuda_12",
  "customTemplateId": "<string>",
  "primePortMapping": [
    {
      "internal": "8080",
      "external": "1234",
      "protocol": "TCP",
      "usedBy": "SSH",
      "description": "<string>"
    }
  ],
  "sshConnection": "<string>",
  "ip": "<string>",
  "resources": {
    "memory": "128",
    "disk": "1000",
    "vcpus": "32"
  },
  "attachedResources": [
    {
      "resourceType": "DISK",
      "id": "<string>",
      "status": "UNATTACHED",
      "isDetachable": true,
      "mountPath": "/workspace",
      "resourcePath": "<string>",
      "size": "<string>"
    }
  ],
  "isSpot": true,
  "autoRestart": true
}

Authorizations

Authorization
string
header
required

Bearer authentication header of the form Bearer <token>, where <token> is your auth token.

Body

application/json
pod
object
required
provider
object
required
team
object | null

Response

200
application/json
Successful Response
userId
string
required

ID of the user associated with this pod, if applicable.

name
string
required

Name of the pod.

type
enum<string>
required

Type of the pod, based on PodTypeEnum.

Available options:
HOSTED,
EXTERNAL
providerType
enum<string>
required

Type of provider associated with the pod, based on ProviderTypeEnum.

Available options:
runpod,
fluidstack,
lambdalabs,
hyperstack,
oblivus,
cudocompute,
scaleway,
tensordock,
datacrunch,
latitude,
crusoecloud,
massedcompute,
akash,
primeintellect,
primecompute,
dc_impala,
dc_kudu,
dc_roan
gpuName
enum<string>
required

Model of the GPU allocated.

Available options:
CPU_NODE,
A10_24GB,
A100_80GB,
A100_40GB,
A30_24GB,
A40_48GB,
RTX3070_8GB,
RTX3070_8GB,
RTX3080_10GB,
RTX3080Ti_12GB,
RTX3090_24GB,
RTX3090Ti_24GB,
RTX4070Ti_12GB,
RTX4080_16GB,
RTX4080Ti_16GB,
RTX4090_24GB,
H100_80GB,
H200_96GB,
H200_141GB,
GH200_480GB,
GH200_624GB,
L4_24GB,
L40_48GB,
L40S_48GB,
RTX4000_8GB,
RTX5000_16GB,
RTX6000_24GB,
RTX8000_48GB,
RTX4000Ada_20GB,
RTX5000Ada_32GB,
RTX6000Ada_48GB,
A2000_6GB,
A4000_16GB,
A4500_20GB,
A5000_24GB,
A6000_48GB,
V100_16GB,
V100_32GB,
P100_16GB,
T4_16GB,
P4_8GB,
P40_24GB
gpuCount
integer
required

Number of GPUs allocated to the node.

Example:

1

socket
enum<string>
required

Type of socket used by the GPU.

Available options:
PCIe,
SXM2,
SXM3,
SXM4,
SXM5
priceHr
number
required

Hourly price for running the pod.

Example:

1.23

id
string

Unique identifier for the pod, generated as a UUID.

teamId
string | null

ID of the team owning this pod, if applicable.

walletId
string | null

ID of the wallet associated with this pod for billing or resource tracking.

status
enum<string>

Current status of the pod, based on PodStatusEnum.

Available options:
PROVISIONING,
PENDING,
ACTIVE,
STOPPED,
ERROR,
TERMINATED
installationStatus
enum<string>

Installation status of the pod, based on InstallationStatusEnum.

Available options:
PENDING,
ACTIVE,
FINISHED,
ERROR,
TERMINATED
installationFailure
string | null

Details about any installation failures that occurred, if applicable.

installationProgress
integer | null

Percentage of the installation process completed.

createdAt
string

Timestamp when the pod was created.

updatedAt
string

Timestamp when the pod was last updated.

jupyterPassword
string | null

Password for accessing the Jupyter environment on the pod, if applicable.

stoppedPriceHr
number | null

Hourly price when the pod is stopped. If empty then priceHr is used.

Example:

0.005

provisioningPriceHr
number | null

Hourly price during the provisioning process. If empty then priceHr is used.

basePriceHr
number | null

Base hourly price for the pod. If the base currency is set.

baseCurrency
string | null

Currency in which the base price is calculated.

environmentType
enum<string>

Type of image selected for the pod.

Available options:
ubuntu_22_cuda_12,
cuda_12_1_pytorch_2_2,
cuda_11_8_pytorch_2_1,
cuda_12_1_pytorch_2_3,
cuda_12_1_pytorch_2_4,
cuda_12_4_pytorch_2_4,
cuda_12_4_pytorch_2_5,
stable_diffusion,
axolotl,
bittensor,
hivemind,
petals_llama,
vllm_llama_8b,
vllm_llama_70b,
vllm_llama_405b,
custom_template,
flux
customTemplateId
string | null

ID of the custom template applied to the pod, if applicable.

primePortMapping
object[] | null

Port mapping.

sshConnection

SSH connection/connections details.

ip

IP address/addresses of the instance.

resources
object | null

Instance resources.

attachedResources
object[] | null

Instance attached resources.

isSpot
boolean | null

Whether the instance is spot.

autoRestart
boolean | null

Automatically restart the instance.