From fa86f4f91ecf20e2deab54985875979cbc9260c5 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?F=C3=A9lix-Antoine=20Fortin?= Date: Wed, 15 Jan 2025 11:38:42 -0500 Subject: [PATCH] Update gcp example --- docs/README.md | 3 ++- examples/gcp/main.tf | 2 +- 2 files changed, 3 insertions(+), 2 deletions(-) diff --git a/docs/README.md b/docs/README.md index a29da73a..4eed92b4 100644 --- a/docs/README.md +++ b/docs/README.md @@ -548,7 +548,8 @@ the operating system and service software and with x86-64 processors (see [NVIDIA/mig-parted issue #30](https://github.com/NVIDIA/mig-parted/issues/30)). 6. `shard`: total number of [Sharding](https://slurm.schedmd.com/gres.html#Sharding) on the node. Sharding allows sharing the same GPU on multiple jobs. The total number of shards is evenly distributed across all GPUs on the node. -The instance specifications are retrieved from the cloud provider data source, but it is possible to explictly specify them. +The instance specifications are retrieved from the cloud provider data source, but it is possible to explicitly specify them. + 7. `cpus`: number of logical processors on the node - [CPUs in slurm.conf](https://slurm.schedmd.com/slurm.conf.html#OPT_CPUs) 8. `ram`: size of real memory on the node in megabyte - [RealMemory in slurm.conf](https://slurm.schedmd.com/slurm.conf.html#OPT_RealMemory) 9. `gpus`: number of graphical processor on the node - [Gres=gpu: in slurm.conf](https://slurm.schedmd.com/slurm.conf.html#OPT_Gres_1) diff --git a/examples/gcp/main.tf b/examples/gcp/main.tf index df15007a..2b2ca4ac 100644 --- a/examples/gcp/main.tf +++ b/examples/gcp/main.tf @@ -24,8 +24,8 @@ module "gcp" { type = "n1-standard-2", tags = ["node"], count = 1, + gpus = 1 gpu_type = "nvidia-tesla-t4", - gpu_count = 1 } }