Credits usage
Depending on your computer vision workflows, the amount of computing power you need will vary. As a general rule, the more computing power your deployments require, the more credits they consume.
To help you choose the right plan, the credits usage for each compute infrastructure is shown below.
Serverless deployments
For serverless deployments (CPU only), you are charged based on the workflow execution time in seconds. You are only charged when you call the deployment endpoint. See infrastructure specifications.
Size | vCPU | RAM | Credits/s |
---|---|---|---|
S | 3 | 4GB | 0.829 |
M | 4 | 6GB | 0.836 |
L | 5 | 8GB | 0.842 |
XL | 6 | 10GB | 0.849 |
Optimizing serverless costs
Due to their pricing model, choosing the cheapest serverless deployment (lowest compute power) is not always the best financial choice.
As an example of credits usage for a concrete workflow, we propose a simulation with a classical OCR workflow:
- basic pre-processing algorithms (noise reduction and luminosity correction)
- text detection with MMLAB framework
- text recognition with MMLAB framework
This table shows the number of images that can be processed with this workflow and the monthly version of the plans:
Size | Execution time (s) | Credits/image | Starter plan (images/month) | Basic plan (images/month) | Pro plan (images/month) |
---|---|---|---|---|---|
S | 16 | 13.26 | 151 | 754 | 6031 |
M | 12 | 10.03 | 199 | 997 | 7974 |
L | 9.6 | 8.08 | 247 | 1237 | 9897 |
XL | 8 | 6.79 | 294 | 1472 | 11779 |
As you can see, in this case, choosing a more powerful deployment is sufficiently beneficial in terms of execution time to offset the higher costs associated with it. To optimize costs, it is crucial to find the right trade-off between compute power and execution time, this may depend on the specific implementation of your workflow algorithms.
CPU instances
For CPU instance deployment, you are charged based on the time the instance is running (in seconds). See infrastructure specifications.
This table shows the lifetime of one deployment for the monthly version of the plans:
Size | Provider | vCPU | RAM | Credits/s | Starter plan | Basic plan | Pro plan |
---|---|---|---|---|---|---|---|
XS | AWS | 2 | 8GB | 0.00477 | ~116h (4.9 days) | ~582h (24.3 days) | ~4659h (194.1 days) |
XS | GCP | 2 | 8GB | 0.00497 | ~112h (4.7 days) | ~559h (23.3 days) | ~4471h (186.3 days) |
XS | Scaleway | 2 | 8GB | 0.00259 | ~215h (8.9 days) | ~1073h (44.7 days) | ~8580h (357.5 days) |
S | AWS | 4 | 8GB | 0.00802 | ~69h (2.9 days) | ~346h (14.4 days) | ~2771h (115.5 days) |
S | GCP | 4 | 16GB | 0.00904 | ~61h (2.6 days) | ~307h (12.8 days) | ~2458h (102.4 days) |
S | Scaleway | 4 | 16GB | 0.00489 | ~114h (4.7 days) | ~568h (23.7 days) | ~4544h (189.4 days) |
M | AWS | 8 | 16GB | 0.01532 | ~36h (1.5 days) | ~181h (7.6 days) | ~1451h (60.4 days) |
M | GCP | 8 | 32GB | 0.01718 | ~32h (1.3 days) | ~162h (6.7 days) | ~1293h (53.9 days) |
M | Scaleway | 8 | 32GB | 0.00943 | ~59h (2.5 days) | ~295h (12.3 days) | ~2357h (98.2 days) |
L | AWS | 16 | 32GB | 0.02991 | ~19h (0.8 days) | ~93h (3.9 days) | ~743h (31 days) |
L | GCP | 16 | 64GB | 0.03345 | ~17h (0.7 days) | ~83h (3.5 days) | ~664h (27.7 days) |
L | Scaleway | 16 | 64GB | 0.01855 | ~30h (1.2 days) | ~150h (6.2 days) | ~1198h (49.9 days) |
XL | AWS | 32 | 64GB | 0.05908 | ~9h (0.4 days) | ~47h (2 days) | ~376h (15.7 days) |
XL | GCP | 32 | 128GB | 0.06599 | ~8h (0.4 days) | ~42h (1.8 days) | ~337h (14 days) |
XL | Scaleway | 32 | 128GB | 0.03684 | ~15h (0.6 days) | ~75h (3.1 days) | ~603h (25.1 days) |
GPU instances
For GPU instance deployment, you are charged based on the time the instance is running (in seconds). See infrastructure specifications.
This table shows the lifetime of one deployment for the monthly version of the plans:
Size | Provider | vCPU | RAM | GPU | Credits/s | Starter plan | Basic plan | Pro plan |
---|---|---|---|---|---|---|---|---|
XS | AWS | 4 | 16GB | NVIDIA T4 16GB | 0.02132 | ~26h (1.1 days) | ~130h (5.4 days) | ~1042h (43.4 days) |
XS | GCP | 4 | 16GB | NVIDIA L4 24GB | 0.02687 | ~21h (0.9 days) | ~103h (4.3 days) | ~827h (34.5 days) |
XS | Scaleway | 8 | 16GB | NVIDIA RTX 3070 8GB | 0.03634 | ~15h (0.6 days) | ~76h (3.2 days) | ~612h (25.5 days) |
S | AWS | 8 | 32GB | NVIDIA T4 16GB | 0.03263 | ~17h (0.7 days) | ~85h (3.5 days) | ~681h (28.4 days) |
S | GCP | 8 | 32GB | NVIDIA L4 24GB | 0.03495 | ~16h (0.7 days) | ~79h (3.3 days) | ~636h (26.5 days) |
S | Scaleway | 8 | 48GB | NVIDIA L4 24GB | 0.03235 | ~17h (0.7 days) | ~86h (3.6 days) | ~687h (28.6 days) |
M | AWS | 4 | 16GB | NVIDIA A10 24GB | 0.03826 | ~15h (0.6 days) | ~73h (3 days) | ~581h (24.2 days) |
M | GCP | 16 | 64GB | NVIDIA L4 24GB | 0.04662 | ~12h (0.5 days) | ~60h (2.5 days) | ~477h (19.9 days) |
M | Scaleway | 16 | 96GB | 2x NVIDIA L4 24GB | 0.0636 | ~9h (0.4 days) | ~44h (1.8 days) | ~349h (14.6 days) |
L | AWS | 8 | 32GB | NVIDIA A10 24GB | 0.04975 | ~11h (0.5 days) | ~56h (2.3 days) | ~447h (18.6 days) |
L | GCP | 12 | 85GB | NVIDIA A100 40GB | 0.12586 | ~4h (0.2 days) | ~22h (0.9 days) | ~177h (7.4 days) |
L | Scaleway | 8 | 96GB | NVIDIA L40S 48GB | 0.05151 | ~11h (0.4 days) | ~54h (2.2 days) | ~431h (18 days) |
XL | AWS | 16 | 64GB | NVIDIA A10 24GB | 0.06636 | ~8h (0.3 days) | ~42h (1.7 days) | ~335h (14 days) |
XL | GCP | 12 | 170GB | NVIDIA A100 80GB | 0.17426 | ~3h (0.1 days) | ~16h (0.7 days) | ~128h (5.3 days) |
XL | Scaleway | 24 | 240GB | NVIDIA H100 80GB | 0.11485 | ~5h (0.2 days) | ~24h (1 days) | ~193h (8.1 days) |
What happens when credits run out
When your credits run out, you will no longer be able to deploy new workflows. You will have to wait for your next refund date depending on whether you are on a monthly or yearly plan. You can also upgrade your plan to get more credits. You will receive your new credits immediately after subscription validation (see plans and pricing).
For active deployments, the behaviour depends on the compute infrastructure:
- Serverless: Deployments are being preserved but you are not able to send requests on them.
- CPU/GPU instances: Deployments (and the infrastructure behind them) are deleted and the endpoint URL becomes invalid. Workflows are being preserved so that you can then deploy when your account is credited.
We will send notification emails to individual user or organization owners when credits are low. You should then make the right decision for your active deployments.