Serverless Model Deployment

Serverless Model Deployment

Serverless Model Deployment

Start running your AI models instantly with Datawizz - and only pay for the tokens you use. No hourly costs, no infrastructure to manage and no scaling to worry about.

Per-Token Pricing

Per-Token Pricing

Per-Token Pricing

Pay only for what you use

Traditional model deployments make you commit to servers and pay hourly, used or not. With Datawizz, you pay per token, and we handle scaling based on your usage.

No complex financial planning or risky commitments—just simple, per-token pricing like OpenAI (well, a lot cheaper though).

API Requests

Number of API requests made to the project

Active projects

Proxy project

Llama 3.2 1B

$

455.75

563,75

3,038M

3,753M

tokens

Phoenix project

Ministral 8B

$

125.55

134.30

8,370M

8,953M

tokens

API Requests

Number of API requests made to the project

Active projects

Proxy project

Llama 3.2 1B

$

455.75

563,75

3,038M

3,753M

tokens

Phoenix project

Ministral 8B

$

125.55

134.30

8,370M

8,953M

tokens

API Requests

Number of API requests made to the project

Active projects

Proxy project

Llama 3.2 1B

$

455.75

563,75

3,038M

3,753M

tokens

Phoenix project

Ministral 8B

$

125.55

134.30

8,370M

8,953M

tokens

Edge Deployment

Edge Deployment

Edge Deployment

Reduce latency with Cloudflare

Supported models are deployed on Cloudflare’s edge network - putting your AI closer to your users and reducing inference latency. Give your users a faster experience with edge deployed models.

Current token usage

Track your token usage in real-time

Active projects

Proxy project

Llama 3.2 1B

873K

958K

/1M

126 832

41 866

tokens left

Phoenix project

Ministral 8B

417K

713K

/1M

582 433

286 544

tokens left

Current token usage

Track your token usage in real-time

Active projects

Proxy project

Llama 3.2 1B

873K

958K

/1M

126 832

41 866

tokens left

Phoenix project

Ministral 8B

417K

713K

/1M

582 433

286 544

tokens left

Current token usage

Track your token usage in real-time

Active projects

Proxy project

Llama 3.2 1B

873K

958K

/1M

126 832

41 866

tokens left

Phoenix project

Ministral 8B

417K

713K

/1M

582 433

286 544

tokens left

Scale with Ease

Scale with Ease

Scale with Ease

Grow your model 

without limits

Datawizz scales up automatically with your usage - no need to worry about server scaling, GPU waitlists and complex infrastructure deployments. Deploy a model in under 60 seconds and forget about it.

Pricing

Pricing

Pricing

How cheap it really is

Pricing depends on your selected base model, and applies for the public models, as well as any models you train from it.

Pricing depends on your selected base model, and applies for the public models, as well as any models you train from it.

Pricing depends on your selected base model, and applies for the public models, as well as any models you train from it.

Base Model

Input Tokens ($/1M)

Output Tokens ($/1M)

Llama 3.2 1B

$0.10

$0.10

Llama 3.2 3B

$0.15

$0.15

Phi-3 Mini

$0.15

$0.15

Command-R 7B

$0.25

$0.25

Ministral 8B

$0.25

$0.25

Llama 3.3 70B

$1.20

$1.20

Base Model

Input Tokens ($/1M)

Output Tokens ($/1M)

Llama 3.2 1B

$0.10

$0.10

Llama 3.2 3B

$0.15

$0.15

Phi-3 Mini

$0.15

$0.15

Command-R 7B

$0.25

$0.25

Ministral 8B

$0.25

$0.25

Llama 3.3 70B

$1.20

$1.20

Input Tokens

Output Tokens

Base Model

Tokens ($/1M)

Llama 3.2 1B

$0.10

Llama 3.2 3B

$0.15

Phi-3 Mini

$0.15

Command-R 7B

$0.25

Ministral 8B

$0.25

Llama 3.3 70B

$1.20