Pricing designed for your compute needs

  • our-cloud-icon

    Anyscale in our cloud

    Build and deploy your AI apps on our infrastructure.

  • your-cloud-icon

    Anyscale in your cloud

    Build and deploy your AI apps in your own infrastucture.

Usage CalculatorEstimated cost: $0.00/month

Dedicated computeCompute running in Anyscale's Cloud provisioned for you.

Usage (ACs): 0.00
Pricing/month:$0.00

LLM APIsAnyscale hosted LLM endpoints for popular and fine tuned open source models.

  • Foundational model inference
  • Fine Tuned model inference
  • Fine tuning jobs
Price:$0.00

Supported models and instance types

LLM API

Foundational Model Inference

ModelPrice ($/M tokens)
Mistral-7B-Instruct-v0.1$0.15
Llama-2-7b-chat-hf$0.15
gemma-7b-it$0.15
NeuralHermes-2.5-Mistral-7B$0.15
Llama-2-13b-chat-hf$0.25
Mixtral-8x7B-Instruct-v0.1$0.50
Llama-2-70b-chat-hf$1.00
CodeLlama-70b-Instruct-hf$1.00
thenlper-gte-large$0.05
BAAI/bge-large-en-v1.5$0.05

Fine Tuned Model Inference

ModelPrice ($/M tokens)
Llama-2-7b-chat-hf*$0.25
Llama-2-13b-chat-hf$0.50
Llama-2-70b-chat-hf$2.00
mistralai/Mistral-7B-Instruct-v0.1$0.25
mistralai/Mixtral-8x7B-Instruct-v0.1$1.00

Fine Tuning Jobs

ModelPrice ($/M tokens)Fixed Cost/Run
Llama-2-13b-chat-hf$2.00$5.00
Llama-2-70b-chat-hf$4.00$5.00
mistralai/Mistral-7B-Instruct-v0.1$1.00$5.00
mistralai/Mixtral-8x7B-Instruct-v0.1$4.00$5.00

Dedicated compute

ProviderThe provider of the underlying infrastructure.
Select any
ConfigurationThe types of resources and GPUs available on the machine type.
Select any

CPU Only

ProviderNamevCPUsMemoryGPUsAC/hr
AWSm5.2xlarge832-0.5130
m5.4xlarge1664-1.0395
m5.8xlarge32128-2.0790
m5.12xlarge48192-3.1050
m5.16xlarge64256-4.1445
m5.24xlarge96384-6.2235
GCPn2-standard-8832-0.6477
n2-standard-161664-1.2953
n2-standard-3232128-2.5907
n2-standard-4848192-3.8860
n2-standard-6464256-5.1813
n2-standard-9696384-7.7721

NVIDIA A100 80GB

ProviderNamevCPUsMemoryGPUsAC/hr
GCPa2-ultragpu-1g-nvidia-a100-80gb-11217016.7892
a2-ultragpu-2g-nvidia-a100-80gb-224340213.5783
a2-ultragpu-4g-nvidia-a100-80gb-448680427.1566
a2-ultragpu-8g-nvidia-a100-80gb-8961360854.3132

NVIDIA A10G

ProviderNamevCPUsMemoryGPUsAC/hr
AWSg5.2xlarge83211.6335
g5.4xlarge166412.1870
g5.8xlarge3212813.3075
g5.12xlarge4819247.6545
g5.48xlarge192768821.9915

NVIDIA H100 80GB

ProviderNamevCPUsMemoryGPUsAC/hr
GCPa3-highgpu-8g-nvidia-h100-80gb-820818308118.6110

NVIDIA L4

ProviderNamevCPUsMemoryGPUsAC/hr
GCPg2-standard-8-nvidia-l4-183212.0615
g2-standard-16-nvidia-l4-1166412.5105
g2-standard-32-nvidia-l4-13212813.4083
g2-standard-48-nvidia-l4-44819249.1437
g2-standard-96-nvidia-l4-896384818.0820

NVIDIA T4

ProviderNamevCPUsMemoryGPUsAC/hr
AWSg4dn.2xlarge83211.0125
g4dn.4xlarge166411.6200
g4dn.8xlarge3212812.9430
g4dn.16xlarge6425615.8725
g4dn.12xlarge4819245.2785
GCPn1-standard-8-nvidia-t4-16gb-183011.0976
n1-standard-16-nvidia-t4-16gb-1166011.7226
n1-standard-32-nvidia-t4-16gb-13212012.9727
n1-highmem-32-nvidia-t4-16gb-13220813.5857
n1-standard-32-nvidia-t4-16gb-43212044.3902

NVIDIA Tesla V100 16GB

ProviderNamevCPUsMemoryGPUsAC/hr
AWSp3.2xlarge86114.1310
p3.8xlarge32244416.5240
p3.16xlarge64488833.0480
GCPn1-highmem-8-nvidia-v100-16gb-185214.1263
n1-highmem-32-nvidia-v100-16gb-432208416.5052
n1-highmem-64-nvidia-v100-16gb-864416833.0105

Interested in trying out Anyscale?

FAQs

What is an Anyscale Credit?

Can I deploy Anyscale in my own cloud accounts?

Can I leverage existing discounts or commitments with my cloud providers?

Do you support on-premise machines?

What is a Token?

How do I get charged if I use less than 1 Million tokens?