|
1 |
| -<img src='https://s3-us-west-2.amazonaws.com/cortex-public/logo.png' height='42'> |
2 |
| - |
3 |
| -<br> |
4 |
| - |
5 | 1 | [Website](https://www.cortex.dev) • [Slack](https://community.cortex.dev) • [Docs](https://docs.cortex.dev)
|
6 | 2 |
|
7 |
| -# Cost-effective serverless computing at scale |
| 3 | +<br> |
8 | 4 |
|
9 |
| -Cortex is a serverless platform for compute-intensive applications. |
| 5 | +<img src='https://s3-us-west-2.amazonaws.com/cortex-public/logo.png' height='42'> |
10 | 6 |
|
11 | 7 | <br>
|
12 | 8 |
|
13 |
| -## Use cases |
| 9 | +# Scale compute-intensive serverless workloads |
14 | 10 |
|
15 |
| -* **Data processing** - run image processing, natural language processing, and more. |
16 |
| -* **Machine learning in production** - train and serve machine learning models in production. |
17 |
| -* **Large-scale inference** - efficiently scale realtime and batch inference workloads. |
| 11 | +One platform for all your microservices, data processing, and machine learning workloads. |
18 | 12 |
|
19 | 13 | <br>
|
20 | 14 |
|
21 |
| -## Scalable |
22 |
| - |
23 |
| -* **Cluster autoscaling** - configure Cortex to spin up instances when load increases and spin them down when load decreases. |
24 |
| -* **Workload autoscaling** - customize the autoscaling behavior of each workload to ensure efficient use of cluster resources. |
| 15 | +## Forget about resource limits, cold starts, and timeouts |
25 | 16 |
|
26 |
| -```text |
27 |
| -$ cortex cluster info |
28 |
| -
|
29 |
| -region: us-east-1 |
30 |
| -instances: [c5.xlarge, g4dn.xlarge] |
31 |
| -spot_instances: true |
32 |
| -min_instances: 10 |
33 |
| -max_instances: 100 |
34 |
| -``` |
| 17 | +* **No resource limits** - allocate as much CPU, GPU, and memory as each workload requires. |
| 18 | +* **No cold starts** - keep a minimum number of API replicas running to ensure that requests are handled in real-time. |
| 19 | +* **No timeouts** - run workloads for as long as you want. |
35 | 20 |
|
36 | 21 | <br>
|
37 | 22 |
|
38 |
| -## Flexible |
| 23 | +## Run realtime, batch, and async workloads |
39 | 24 |
|
40 |
| -* **Any workload** - define custom Python functions or containers and deploy them as realtime, async, and batch workloads. |
41 |
| -* **Any pipeline** - chain workloads seamlessly to create custom data pipelines. |
| 25 | +* **Realtime** - realtime APIs respond to requests in real-time and autoscale based on in-flight request volumes. |
| 26 | +* **Batch** - batch APIs run distributed and fault-tolerant batch processing jobs on-demand. |
| 27 | +* **Async** - async APIs process requests asynchronously and autoscale based on request queue length. |
42 | 28 |
|
43 |
| -```text |
44 |
| -$ cortex deploy apis.yaml |
| 29 | +<br> |
45 | 30 |
|
46 |
| -creating text-generator (realtime API) |
47 |
| -creating image-classifier (batch API) |
48 |
| -creating video-analyzer (async API) |
| 31 | +## Minimize costs, especially at scale |
49 | 32 |
|
50 |
| -all APIs are ready! |
51 |
| -``` |
| 33 | +* **Spot instance management** - cortex can automatically run your workloads on spot instances so that you can save money without sacrificing reliability. |
| 34 | +* **Customizable autoscaling** - optimize the autoscaling behavior for each workload to ensure efficient resource utilization. |
52 | 35 |
|
53 | 36 | <br>
|
54 | 37 |
|
55 |
| -## Observable |
56 |
| - |
57 |
| -* **Structured logging** - automatically stream logs from every workload to your favorite log management tool. |
58 |
| -* **Metrics aggregation** - monitor your workloads with pre-built Grafana dashboards and add your own custom dashboards. |
| 38 | +## Streamline your development workflow |
59 | 39 |
|
60 |
| -```text |
61 |
| -$ cortex get |
62 |
| -
|
63 |
| -API TYPE REPLICAS |
64 |
| -text-generator realtime 32 |
65 |
| -image-classifier batch 64 |
66 |
| -video-analyzer async 16 |
67 |
| -``` |
| 40 | +* **Deploy functions and containers** - upload Python functions and let Cortex manage dependencies or connect to your image registry. |
| 41 | +* **Deploy programmatically** - deploy your code interactively with a CLI or programmatically with a Python client. |
| 42 | +* **Debug locally** - keep your local and cloud workflows similar so that you can iterate on your code faster. |
68 | 43 |
|
69 | 44 | <br>
|
70 | 45 |
|
71 |
| -## Affordable |
| 46 | +## Integrate with modern observability tools |
| 47 | + |
| 48 | +* **Stream structured logs** - automatically stream logs from every workload to your favorite log management tool. |
| 49 | +* **Collect custom metrics** - monitor your workloads with pre-built dashboards and add your own custom dashboards. |
72 | 50 |
|
73 |
| -* **Spot instance management** - spend less on EC2 instances by letting Cortex manage spot instances. |
74 |
| -* **Multi-instance type clusters** - configure resources per workload to run each workload on the right hardware. |
| 51 | +<br> |
75 | 52 |
|
76 |
| -```text |
77 |
| -$ cortex cluster pricing |
| 53 | +## Deploy Cortex on your AWS account |
78 | 54 |
|
79 |
| -RESOURCE COST PER HOUR |
80 |
| -1 eks cluster $0.10 |
81 |
| -2 network load balancers $0.02 each |
82 |
| -10-100 g4dn.xlarge instances $0.53 each |
83 |
| -10-100 c5.xlarge instances $0.17 each |
84 |
| -``` |
| 55 | +* **Identity & Access Management** - Cortex integrates with IAM to simplify authentication and authorization worklfows. |
| 56 | +* **Elastic Kubernetes Service** - Cortex runs on top of EKS to scale workloads reliably and cost-effectively. |
| 57 | +* **Billing & Cost Management** - Cortex supports resource tagging to enable detailed spend visibility. |
0 commit comments