Skip to content

Commit

Permalink
Add nanotron performance
Browse files Browse the repository at this point in the history
  • Loading branch information
xrsrke authored Jan 23, 2025
1 parent 9055c66 commit 35766d2
Showing 1 changed file with 9 additions and 0 deletions.
9 changes: 9 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -51,6 +51,7 @@ pip install triton "flash-attn>=2.5.0" --no-build-isolation
> [!TIP]
> We log to wandb automatically if it's installed. For that you can use `pip install wandb`. If you don't want to use wandb, you can run `wandb disabled`.

## Quick Start
### Training a tiny Llama model
The following command will train a tiny Llama model on a single node with 8 GPUs. The model will be saved in the `checkpoints` directory as specified in the config file.
Expand All @@ -64,6 +65,14 @@ torchrun --nproc_per_node=1 run_generate.py --ckpt-path checkpoints/10/ --tp 1 -
# We could set a larger TP for faster generation, and a larger PP in case of very large models.
```

### Performance
| | 1 node | 4 nodes | 8 nodes | 16 nodes |
| --- | --- | --- | --- | --- |
| **8b** | | | | |
| nanotron | 45.22% | 43% | | 36% |
| **80b** | | | | |
| nanotron | | | 30.65% | 34.29% |

### Custom examples
You can find more examples in the [`/examples`](/examples) directory:
<!-- Make a table of the examples we support -->
Expand Down

0 comments on commit 35766d2

Please sign in to comment.