Skip to content

Commit

Permalink
renaming
Browse files Browse the repository at this point in the history
  • Loading branch information
artek0chumak committed Sep 24, 2024
1 parent 0fa00dc commit c71510a
Show file tree
Hide file tree
Showing 3 changed files with 4 additions and 4 deletions.
4 changes: 2 additions & 2 deletions src/together/cli/api/finetune.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@
from tabulate import tabulate

from together import Together
from together.cli.api.utils import AUTO_INT
from together.cli.api.utils import INT_WITH_MAX
from together.utils import finetune_price_to_dollars, log_warn, parse_timestamp
from together.types.finetune import DownloadCheckpointType, FinetuneTrainingLimits

Expand Down Expand Up @@ -46,7 +46,7 @@ def fine_tuning(ctx: click.Context) -> None:
@click.option(
"--n-checkpoints", type=int, default=1, help="Number of checkpoints to save"
)
@click.option("--batch-size", type=AUTO_INT, default="max", help="Train batch size")
@click.option("--batch-size", type=INT_WITH_MAX, default="max", help="Train batch size")
@click.option("--learning-rate", type=float, default=1e-5, help="Learning rate")
@click.option(
"--lora/--no-lora",
Expand Down
2 changes: 1 addition & 1 deletion src/together/cli/api/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,4 +18,4 @@ def convert(
self.fail("Invalid integer value: {value}")


AUTO_INT = AutoIntParamType()
INT_WITH_MAX = AutoIntParamType()
2 changes: 1 addition & 1 deletion src/together/resources/finetune.py
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@ def create(
n_evals (int, optional): Number of evaluation loops to run. Defaults to 0.
n_checkpoints (int, optional): Number of checkpoints to save during fine-tuning.
Defaults to 1.
batch_size (int, optional): Batch size for fine-tuning. Defaults to auto.
batch_size (int, optional): Batch size for fine-tuning. Defaults to max.
learning_rate (float, optional): Learning rate multiplier to use for training
Defaults to 0.00001.
lora (bool, optional): Whether to use LoRA adapters. Defaults to True.
Expand Down

0 comments on commit c71510a

Please sign in to comment.