Skip to content

Commit

Permalink
Merge branch 'main' into litgpt-eval
Browse files Browse the repository at this point in the history
  • Loading branch information
rasbt authored Apr 1, 2024
2 parents a881630 + 449eb29 commit 1ca218b
Show file tree
Hide file tree
Showing 5 changed files with 17 additions and 7 deletions.
4 changes: 2 additions & 2 deletions litgpt/finetune/adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,8 +46,8 @@ def setup(
train: TrainArgs = TrainArgs(
save_interval=1000,
log_interval=1,
global_batch_size=128,
micro_batch_size=4,
global_batch_size=16,
micro_batch_size=1,
lr_warmup_steps=100,
epochs=5,
learning_rate=1e-3,
Expand Down
4 changes: 2 additions & 2 deletions litgpt/finetune/adapter_v2.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,8 +46,8 @@ def setup(
train: TrainArgs = TrainArgs(
save_interval=1000,
log_interval=1,
global_batch_size=128,
micro_batch_size=4,
global_batch_size=16,
micro_batch_size=1,
lr_warmup_steps=100,
epochs=5,
learning_rate=1e-3,
Expand Down
2 changes: 1 addition & 1 deletion litgpt/finetune/full.py
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@ def setup(
train: TrainArgs = TrainArgs(
save_interval=1000,
log_interval=1,
global_batch_size=64,
global_batch_size=16,
micro_batch_size=1,
lr_warmup_steps=100,
epochs=5,
Expand Down
4 changes: 2 additions & 2 deletions litgpt/finetune/lora.py
Original file line number Diff line number Diff line change
Expand Up @@ -56,8 +56,8 @@ def setup(
train: TrainArgs = TrainArgs(
save_interval=1000,
log_interval=1,
global_batch_size=128,
micro_batch_size=4,
global_batch_size=16,
micro_batch_size=1,
lr_warmup_steps=100,
epochs=5,
learning_rate=3e-4,
Expand Down
10 changes: 10 additions & 0 deletions tutorials/download_model_weights.md
Original file line number Diff line number Diff line change
Expand Up @@ -154,6 +154,16 @@ unsloth/Mistral-7B-v0.2
> [!TIP]
> To sort the list above by model name after the `/`, use `litgpt download | sort -f -t'/' -k2`.
 

> [!NOTE]
> If you want to adopt a model variant that is not listed in the table above but has a similar architecture as one of the supported models, you can use this model by by using the `--model_name` argument as shown below:
> ```bash
> litgpt download \
> --repo_id NousResearch/Hermes-2-Pro-Mistral-7B \
> --model_name Mistral-7B-v0.1
> ```
 
### 2. Download Model Weights
Expand Down

0 comments on commit 1ca218b

Please sign in to comment.