From 4251b2752e9cc00d1aa90eb3bb751c28f237182d Mon Sep 17 00:00:00 2001 From: Arsh Zahed Date: Mon, 18 Nov 2024 15:51:44 -0800 Subject: [PATCH] Fix more typing --- src/together/resources/finetune.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/src/together/resources/finetune.py b/src/together/resources/finetune.py index a5b1ce83..ceb7bf0b 100644 --- a/src/together/resources/finetune.py +++ b/src/together/resources/finetune.py @@ -37,8 +37,8 @@ def createFinetuneRequest( n_checkpoints: int | None = 1, batch_size: int | Literal["max"] = "max", learning_rate: float | None = 0.00001, - min_lr_ratio: float | None = 0.0, - warmup_ratio: float | None = 0.0, + min_lr_ratio: float = 0.0, + warmup_ratio: float = 0.0, max_grad_norm: float = 1.0, weight_decay: float = 0.0, lora: bool = False, @@ -140,7 +140,7 @@ def create( batch_size: int | Literal["max"] = "max", learning_rate: float | None = 0.00001, min_lr_ratio: float = 0.0, - warmup_ratio: float | None = 0.0, + warmup_ratio: float = 0.0, max_grad_norm: float = 1.0, weight_decay: float = 0.0, lora: bool = False, @@ -469,7 +469,7 @@ async def create( batch_size: int | Literal["max"] = "max", learning_rate: float | None = 0.00001, min_lr_ratio: float = 0.0, - warmup_ratio: float | None = 0.0, + warmup_ratio: float = 0.0, max_grad_norm: float = 1.0, weight_decay: float = 0.0, lora: bool = False,