Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion psyche-book/src/development/contributing.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ It's not a deal-breaker, but rebase makes us happy \<3
### Clean Linear History

Rebasing creates a linear commit history without merges going back and forth, making it much easier to identify the place a change was made.
Fix-ups in merge commits that introduce bugs are no longer associated with the original code, whereas with with rebase you'd find the bug as part of its original commit.
Fix-ups in merge commits that introduce bugs are no longer associated with the original code, whereas with rebase you'd find the bug as part of its original commit.

Merge commits add extra noise to the history without adding meaningful content about what changed.

Expand Down
2 changes: 1 addition & 1 deletion psyche-book/src/development/python.md
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ When you use the `--python` flag, Psyche automatically spawns Python sidecar pro
python -m psyche.sidecar --parent-pid <pid> --backend <backend> --init-method <method> --world-size <size> --rank <rank>
```

By default only one sidecar using one GPU will be spawned, the amount will change depending on two different arguments `--data-parallelism` and `--tensor-parallelism`. The first one will spawned one entire copy of the model per GPU and the latter will split the model across multiple GPUs. The amount of sidecars spawned will be the product of these two arguments. Take into account that you will need `tensor_parallelism * data_parallelism` GPUs to run that amount of sidecars.
By default only one sidecar using one GPU will be spawned, the amount will change depending on two different arguments `--data-parallelism` and `--tensor-parallelism`. The first one will spawn one entire copy of the model per GPU and the latter will split the model across multiple GPUs. The amount of sidecars spawned will be the product of these two arguments. Take into account that you will need `tensor_parallelism * data_parallelism` GPUs to run that amount of sidecars.

Here's an overview of the different options that the `psyche-sidecar` provides in case you want to test sidecars with different configurations.

Expand Down
4 changes: 2 additions & 2 deletions psyche-book/src/development/running-offchain.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ Since we want to run many clients and the server we'll need several terminal win
### Running

Since the local-testnet examples uses a local server to provide the data for the clients to train on, you'll need to download the data first.
The best way to do it is install the HuggingFace CLI tool running `curl -LsSf https://hf.co/cli/install.sh | bash`, once installed just run the following command to get some random data and place it in the correct place for the local server to use it:
The best way to do it is to install the HuggingFace CLI tool by running `curl -LsSf https://hf.co/cli/install.sh | bash`. Once installed, run the following command to get some random data and place it in the correct place for the local server to use it:

```bash
hf download emozilla/fineweb-10bt-tokenized-datatrove-llama2 --repo-type dataset --local-dir ./data/fineweb-10bt
Expand All @@ -37,7 +37,7 @@ just local-testnet \

This will run a server locally that acts as the coordinator and 3 clients that will connect to the server and start training on the downloaded data. We'll talk about the configuration of the run later on but this example will use the config located at `./config/consilience-match-llama2-20m-fineweb-pretrain-dev/state.toml`, there you can have a glimpse of the configuration options.

There's a _lot_ of options to configure the local testnet. Check em out below to configure runs as you see fit:
There's a _lot_ of options to configure the local testnet. Check them out below to configure runs as you see fit:

<details>
<summary>Command-line options</summary>
Expand Down
3 changes: 1 addition & 2 deletions psyche-book/src/development/setup.md
Original file line number Diff line number Diff line change
Expand Up @@ -15,8 +15,7 @@ To install Nix, simply run the `./setup-nix.sh` script. This will install Nix an

##### Binary cache

If you already have Nix installed, or are installing it manually,
To speed up your builds & your local dev shell, we recommend enabling the binary cache from `garnix`, our CI provider.
To speed up your builds and your local dev shell, we recommend enabling the binary cache from `garnix`, our CI provider.

In order to use the cache that garnix provides, change your `nix.conf`, adding `https://cache.garnix.io` to substituters, and `cache.garnix.io:CTFPyKSLcx5RMJKfLo5EEPUObbA78b0YQ2DTCJXqr9g=` to `trusted-public-keys`.

Expand Down