Skip to content

Commit 46a3687

Browse files
[Blog] How Toffee streamlines inference and cut GPU costs with dstack
Added thumbnail
1 parent 61b6711 commit 46a3687

1 file changed

Lines changed: 2 additions & 2 deletions

File tree

docs/blog/posts/toffee.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ title: "How Toffee streamlines inference and cut GPU costs with dstack"
33
date: 2025-12-05
44
description: "TBA"
55
slug: toffee
6-
image: https://dstack.ai/static-assets/static-assets/images/dstack-ea-slide-2-background-min.png
6+
image: https://dstack.ai/static-assets/static-assets/images/dstack-toffee.png
77
categories:
88
- Case studies
99
links:
@@ -14,7 +14,7 @@ links:
1414

1515
In a recent engineering [blog post](https://research.toffee.ai/blog/how-we-use-dstack-at-toffee), Toffee shared how they use `dstack` to run large-language and image-generation models across multiple GPU clouds, while keeping their core backend on AWS. This case study summarizes key insights and highlights how `dstack` became the backbone of Toffee’s multi-cloud inference stack.
1616

17-
<img src="https://dstack.ai/static-assets/static-assets/images/dstack-ea-slide-1.png" width="630" />
17+
<img src="https://dstack.ai/static-assets/static-assets/images/dstack-toffee.png" width="630" />
1818

1919
<!-- more -->
2020

0 commit comments

Comments
 (0)