You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -41,7 +41,7 @@ for AI workloads both in the cloud and on-prem, speeding up the development, tra
41
41
To use `dstack` with your own cloud accounts, create the `~/.dstack/server/config.yml` file and
42
42
[configure backends](https://dstack.ai/docs/reference/server/config.yml). Alternatively, you can configure backends via the control plane UI after you start the server.
43
43
44
-
You can skip backends configuration if you intend to run containers only on your on-prem servers. Use [SSH fleets](https://dstack.ai/docs/concepts/fleets#ssh-fleets) for that.
44
+
You can skip backends configuration if you intend to run containers only on your on-prem servers. Use [SSH fleets](https://dstack.ai/docs/concepts/fleets#ssh) for that.
Copy file name to clipboardExpand all lines: docs/blog/posts/amd-mi300x-inference-benchmark.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,7 +11,7 @@ categories:
11
11
12
12
# Benchmarking Llama 3.1 405B on 8x AMD MI300X GPUs
13
13
14
-
At `dstack`, we've been adding support for AMD GPUs with [SSH fleets](../../docs/concepts/fleets.md#ssh-fleets),
14
+
At `dstack`, we've been adding support for AMD GPUs with [SSH fleets](../../docs/concepts/fleets.md#ssh),
15
15
so we saw this as a great chance to test our integration by benchmarking AMD GPUs. Our friends at
16
16
[Hot Aisle :material-arrow-top-right-thin:{ .external }](https://hotaisle.xyz/){:target="_blank"}, who build top-tier
17
17
bare metal compute for AMD GPUs, kindly provided the hardware for the benchmark.
@@ -35,7 +35,7 @@ Here is the spec of the bare metal machine we got:
35
35
??? info "Set up an SSH fleet"
36
36
37
37
Hot Aisle provided us with SSH access to the machine. To make it accessible via `dstack`,
38
-
we created an [SSH fleet](../../docs/concepts/fleets.md#ssh-fleets) using the following configuration:
38
+
we created an [SSH fleet](../../docs/concepts/fleets.md#ssh) using the following configuration:
39
39
40
40
<div editor-title="hotaisle.dstack.yml">
41
41
@@ -216,7 +216,7 @@ If you have questions, feedback, or want to help improve the benchmark, please r
216
216
is the primary sponsor of this benchmark, and we are sincerely grateful for their hardware and support.
217
217
218
218
If you'd like to use top-tier bare metal compute with AMD GPUs, we recommend going
219
-
with Hot Aisle. Once you gain access to a cluster, it can be easily accessed via `dstack`'s [SSH fleet](../../docs/concepts/fleets.md#ssh-fleets) easily.
219
+
with Hot Aisle. Once you gain access to a cluster, it can be easily accessed via `dstack`'s [SSH fleet](../../docs/concepts/fleets.md#ssh) easily.
220
220
221
221
### RunPod
222
222
If you’d like to use on-demand compute with AMD GPUs at affordable prices, you can configure `dstack` to
Copy file name to clipboardExpand all lines: docs/blog/posts/amd-on-runpod.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -39,7 +39,7 @@ you can now specify an AMD GPU under `resources`. Below are a few examples.
39
39
## Configuration
40
40
41
41
=== "Service"
42
-
Here's an example of a [service](../../docs/services.md) that deploys
42
+
Here's an example of a [service](../../docs/concepts/services.md) that deploys
43
43
Llama 3.1 70B in FP16 using [TGI :material-arrow-top-right-thin:{ .external }](https://huggingface.co/docs/text-generation-inference/en/installation_amd){:target="_blank"}.
0 commit comments